var/home/core/zuul-output/0000755000175000017500000000000015114310102014511 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114326775015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005622662215114326766017723 0ustar rootrootDec 04 13:48:15 crc systemd[1]: Starting Kubernetes Kubelet... Dec 04 13:48:15 crc restorecon[4701]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:15 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:16 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:16 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:16 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:16 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:16 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:16 crc restorecon[4701]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 13:48:16 crc restorecon[4701]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 04 13:48:16 crc kubenswrapper[4848]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 13:48:16 crc kubenswrapper[4848]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 04 13:48:16 crc kubenswrapper[4848]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 13:48:16 crc kubenswrapper[4848]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 13:48:16 crc kubenswrapper[4848]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 04 13:48:16 crc kubenswrapper[4848]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.187045 4848 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191322 4848 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191361 4848 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191369 4848 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191375 4848 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191381 4848 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191387 4848 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191395 4848 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191405 4848 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191410 4848 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191426 4848 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191434 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191441 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191458 4848 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191467 4848 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191474 4848 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191480 4848 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191485 4848 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191491 4848 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191496 4848 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191501 4848 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191506 4848 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191517 4848 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191521 4848 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191527 4848 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191531 4848 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191536 4848 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191541 4848 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191546 4848 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191551 4848 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191557 4848 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191561 4848 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191567 4848 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191572 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191578 4848 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191587 4848 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191592 4848 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191597 4848 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191603 4848 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191608 4848 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191615 4848 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191621 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191626 4848 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191632 4848 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191637 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191662 4848 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191667 4848 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191677 4848 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191683 4848 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191690 4848 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191696 4848 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191700 4848 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191704 4848 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191708 4848 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191712 4848 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191716 4848 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191720 4848 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191724 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191729 4848 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191736 4848 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191744 4848 feature_gate.go:330] unrecognized feature gate: Example Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191752 4848 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191758 4848 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191764 4848 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191769 4848 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191775 4848 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191781 4848 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191785 4848 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191794 4848 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191799 4848 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191809 4848 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.191815 4848 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192212 4848 flags.go:64] FLAG: --address="0.0.0.0" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192229 4848 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192245 4848 flags.go:64] FLAG: --anonymous-auth="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192259 4848 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192268 4848 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192276 4848 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192287 4848 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192296 4848 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192303 4848 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192311 4848 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192325 4848 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192332 4848 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192340 4848 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192349 4848 flags.go:64] FLAG: --cgroup-root="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192355 4848 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192360 4848 flags.go:64] FLAG: --client-ca-file="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192369 4848 flags.go:64] FLAG: --cloud-config="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192375 4848 flags.go:64] FLAG: --cloud-provider="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192381 4848 flags.go:64] FLAG: --cluster-dns="[]" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192393 4848 flags.go:64] FLAG: --cluster-domain="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192398 4848 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192404 4848 flags.go:64] FLAG: --config-dir="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192410 4848 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192416 4848 flags.go:64] FLAG: --container-log-max-files="5" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192425 4848 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192430 4848 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192436 4848 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192443 4848 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192453 4848 flags.go:64] FLAG: --contention-profiling="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192459 4848 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192465 4848 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192471 4848 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192477 4848 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192485 4848 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192491 4848 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192497 4848 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192506 4848 flags.go:64] FLAG: --enable-load-reader="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192512 4848 flags.go:64] FLAG: --enable-server="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192519 4848 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192528 4848 flags.go:64] FLAG: --event-burst="100" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192533 4848 flags.go:64] FLAG: --event-qps="50" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192540 4848 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192546 4848 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192552 4848 flags.go:64] FLAG: --eviction-hard="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192564 4848 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192570 4848 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192577 4848 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192584 4848 flags.go:64] FLAG: --eviction-soft="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192590 4848 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192603 4848 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192648 4848 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192654 4848 flags.go:64] FLAG: --experimental-mounter-path="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192660 4848 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192671 4848 flags.go:64] FLAG: --fail-swap-on="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192735 4848 flags.go:64] FLAG: --feature-gates="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192745 4848 flags.go:64] FLAG: --file-check-frequency="20s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192752 4848 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192763 4848 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192769 4848 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192775 4848 flags.go:64] FLAG: --healthz-port="10248" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192784 4848 flags.go:64] FLAG: --help="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192790 4848 flags.go:64] FLAG: --hostname-override="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192795 4848 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192801 4848 flags.go:64] FLAG: --http-check-frequency="20s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192808 4848 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192814 4848 flags.go:64] FLAG: --image-credential-provider-config="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192839 4848 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192845 4848 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.192850 4848 flags.go:64] FLAG: --image-service-endpoint="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193480 4848 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193500 4848 flags.go:64] FLAG: --kube-api-burst="100" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193506 4848 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193513 4848 flags.go:64] FLAG: --kube-api-qps="50" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193518 4848 flags.go:64] FLAG: --kube-reserved="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193524 4848 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193530 4848 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193536 4848 flags.go:64] FLAG: --kubelet-cgroups="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193541 4848 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193547 4848 flags.go:64] FLAG: --lock-file="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193553 4848 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193558 4848 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193564 4848 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193611 4848 flags.go:64] FLAG: --log-json-split-stream="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193616 4848 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193621 4848 flags.go:64] FLAG: --log-text-split-stream="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193627 4848 flags.go:64] FLAG: --logging-format="text" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193632 4848 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193638 4848 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193643 4848 flags.go:64] FLAG: --manifest-url="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193648 4848 flags.go:64] FLAG: --manifest-url-header="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193658 4848 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193663 4848 flags.go:64] FLAG: --max-open-files="1000000" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193669 4848 flags.go:64] FLAG: --max-pods="110" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193673 4848 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193678 4848 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193683 4848 flags.go:64] FLAG: --memory-manager-policy="None" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193688 4848 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193695 4848 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193700 4848 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193705 4848 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193732 4848 flags.go:64] FLAG: --node-status-max-images="50" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193739 4848 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193744 4848 flags.go:64] FLAG: --oom-score-adj="-999" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193753 4848 flags.go:64] FLAG: --pod-cidr="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193758 4848 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193770 4848 flags.go:64] FLAG: --pod-manifest-path="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193775 4848 flags.go:64] FLAG: --pod-max-pids="-1" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193780 4848 flags.go:64] FLAG: --pods-per-core="0" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193785 4848 flags.go:64] FLAG: --port="10250" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193790 4848 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193795 4848 flags.go:64] FLAG: --provider-id="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193799 4848 flags.go:64] FLAG: --qos-reserved="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193804 4848 flags.go:64] FLAG: --read-only-port="10255" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193809 4848 flags.go:64] FLAG: --register-node="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193813 4848 flags.go:64] FLAG: --register-schedulable="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193818 4848 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193828 4848 flags.go:64] FLAG: --registry-burst="10" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193833 4848 flags.go:64] FLAG: --registry-qps="5" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193839 4848 flags.go:64] FLAG: --reserved-cpus="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193845 4848 flags.go:64] FLAG: --reserved-memory="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193852 4848 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193858 4848 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193864 4848 flags.go:64] FLAG: --rotate-certificates="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193870 4848 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193876 4848 flags.go:64] FLAG: --runonce="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193881 4848 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193886 4848 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193892 4848 flags.go:64] FLAG: --seccomp-default="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193897 4848 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193902 4848 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193907 4848 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193913 4848 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193918 4848 flags.go:64] FLAG: --storage-driver-password="root" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193923 4848 flags.go:64] FLAG: --storage-driver-secure="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193929 4848 flags.go:64] FLAG: --storage-driver-table="stats" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193934 4848 flags.go:64] FLAG: --storage-driver-user="root" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193939 4848 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193971 4848 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193977 4848 flags.go:64] FLAG: --system-cgroups="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193982 4848 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193994 4848 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.193999 4848 flags.go:64] FLAG: --tls-cert-file="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194004 4848 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194013 4848 flags.go:64] FLAG: --tls-min-version="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194018 4848 flags.go:64] FLAG: --tls-private-key-file="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194023 4848 flags.go:64] FLAG: --topology-manager-policy="none" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194028 4848 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194035 4848 flags.go:64] FLAG: --topology-manager-scope="container" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194040 4848 flags.go:64] FLAG: --v="2" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194048 4848 flags.go:64] FLAG: --version="false" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194055 4848 flags.go:64] FLAG: --vmodule="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194061 4848 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194067 4848 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194265 4848 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194459 4848 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194468 4848 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194474 4848 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194480 4848 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194485 4848 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194492 4848 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194498 4848 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194503 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194508 4848 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194513 4848 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194517 4848 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194526 4848 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194531 4848 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194536 4848 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194540 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194545 4848 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194549 4848 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194553 4848 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194558 4848 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194562 4848 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194567 4848 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194571 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194576 4848 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194580 4848 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194585 4848 feature_gate.go:330] unrecognized feature gate: Example Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194590 4848 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194596 4848 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194601 4848 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194605 4848 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194610 4848 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194614 4848 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194620 4848 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194626 4848 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194631 4848 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194636 4848 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194641 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194646 4848 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194651 4848 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194655 4848 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194660 4848 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194664 4848 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194669 4848 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194673 4848 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194681 4848 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194685 4848 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194690 4848 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194695 4848 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194699 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194703 4848 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194708 4848 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194712 4848 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194716 4848 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194721 4848 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194725 4848 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194729 4848 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194735 4848 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194741 4848 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194747 4848 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194752 4848 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194756 4848 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194761 4848 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194765 4848 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194770 4848 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194774 4848 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194778 4848 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194783 4848 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194787 4848 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194791 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194796 4848 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.194800 4848 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.194808 4848 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.206140 4848 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.206210 4848 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206365 4848 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206385 4848 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206397 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206409 4848 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206420 4848 feature_gate.go:330] unrecognized feature gate: Example Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206431 4848 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206441 4848 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206450 4848 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206460 4848 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206470 4848 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206484 4848 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206501 4848 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206514 4848 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206525 4848 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206537 4848 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206550 4848 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206562 4848 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206574 4848 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206585 4848 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206596 4848 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206606 4848 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206617 4848 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206628 4848 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206641 4848 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206652 4848 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206663 4848 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206673 4848 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206683 4848 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206697 4848 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206711 4848 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206722 4848 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206734 4848 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206743 4848 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206752 4848 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206767 4848 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206778 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206790 4848 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206800 4848 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206812 4848 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206822 4848 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206832 4848 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206842 4848 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206854 4848 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206864 4848 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206872 4848 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206881 4848 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206891 4848 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206901 4848 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206910 4848 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206920 4848 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206931 4848 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206941 4848 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206984 4848 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.206995 4848 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207006 4848 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207017 4848 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207027 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207037 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207047 4848 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207056 4848 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207066 4848 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207076 4848 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207086 4848 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207096 4848 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207106 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207116 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207126 4848 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207136 4848 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207149 4848 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207163 4848 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207176 4848 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.207195 4848 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207587 4848 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207607 4848 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207619 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207630 4848 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207641 4848 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207655 4848 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207668 4848 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207679 4848 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207688 4848 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207697 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207705 4848 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207715 4848 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207725 4848 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207733 4848 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207744 4848 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207753 4848 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207761 4848 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207769 4848 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207777 4848 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207785 4848 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207793 4848 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207800 4848 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207808 4848 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207816 4848 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207827 4848 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207835 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207843 4848 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207852 4848 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207861 4848 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207869 4848 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207878 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207886 4848 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207895 4848 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207903 4848 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207912 4848 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207921 4848 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207928 4848 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207936 4848 feature_gate.go:330] unrecognized feature gate: Example Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207975 4848 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207982 4848 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207991 4848 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.207999 4848 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208007 4848 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208014 4848 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208022 4848 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208030 4848 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208038 4848 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208046 4848 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208054 4848 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208062 4848 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208069 4848 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208080 4848 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208090 4848 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208100 4848 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208108 4848 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208117 4848 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208127 4848 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208135 4848 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208144 4848 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208152 4848 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208161 4848 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208170 4848 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208179 4848 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208188 4848 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208196 4848 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208204 4848 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208212 4848 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208222 4848 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208232 4848 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208241 4848 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.208250 4848 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.208265 4848 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.208984 4848 server.go:940] "Client rotation is on, will bootstrap in background" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.213908 4848 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.214129 4848 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.215169 4848 server.go:997] "Starting client certificate rotation" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.215235 4848 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.215468 4848 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-09 11:40:52.205499207 +0000 UTC Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.215570 4848 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.225564 4848 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.227588 4848 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.229353 4848 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.240195 4848 log.go:25] "Validated CRI v1 runtime API" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.263556 4848 log.go:25] "Validated CRI v1 image API" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.265462 4848 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.268302 4848 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-04-13-43-23-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.268350 4848 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.298671 4848 manager.go:217] Machine: {Timestamp:2025-12-04 13:48:16.296451247 +0000 UTC m=+0.238947845 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:fca2e846-ef30-4639-ace9-fc8f55c29a70 BootID:c51146e5-a7ba-4456-8c4a-822438e48eb1 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:5c:d0:e5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:5c:d0:e5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9a:7e:29 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:53:93:4e Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a0:44:f3 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:86:fd:d5 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:7e:df:8d:e5:8f:9e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a6:5f:d4:27:35:97 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.299267 4848 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.299582 4848 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.300464 4848 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.300778 4848 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.300838 4848 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.301192 4848 topology_manager.go:138] "Creating topology manager with none policy" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.301210 4848 container_manager_linux.go:303] "Creating device plugin manager" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.301529 4848 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.301579 4848 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.302074 4848 state_mem.go:36] "Initialized new in-memory state store" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.302254 4848 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.303810 4848 kubelet.go:418] "Attempting to sync node with API server" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.303853 4848 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.303908 4848 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.303934 4848 kubelet.go:324] "Adding apiserver pod source" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.303988 4848 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.306798 4848 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.306845 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.306907 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.307003 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.307033 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.307388 4848 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.308971 4848 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309737 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309775 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309789 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309802 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309823 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309836 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309852 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309873 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309889 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309904 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309921 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.309934 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.310314 4848 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.310906 4848 server.go:1280] "Started kubelet" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.311639 4848 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.311641 4848 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 04 13:48:16 crc systemd[1]: Started Kubernetes Kubelet. Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.312387 4848 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.312427 4848 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.313333 4848 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.313414 4848 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.313433 4848 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 21:47:30.04666799 +0000 UTC Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.313504 4848 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 871h59m13.733171124s for next certificate rotation Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.316372 4848 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.316426 4848 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.316539 4848 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.317022 4848 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.319215 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="200ms" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.318873 4848 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.5:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e0744dc4a9e56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 13:48:16.310861398 +0000 UTC m=+0.253357966,LastTimestamp:2025-12-04 13:48:16.310861398 +0000 UTC m=+0.253357966,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.321173 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.321305 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.322450 4848 server.go:460] "Adding debug handlers to kubelet server" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.322778 4848 factory.go:153] Registering CRI-O factory Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.322823 4848 factory.go:221] Registration of the crio container factory successfully Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.322999 4848 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.323023 4848 factory.go:55] Registering systemd factory Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.323224 4848 factory.go:221] Registration of the systemd container factory successfully Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.323279 4848 factory.go:103] Registering Raw factory Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.323310 4848 manager.go:1196] Started watching for new ooms in manager Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.326743 4848 manager.go:319] Starting recovery of all containers Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332745 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332813 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332827 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332842 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332854 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332867 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332883 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332897 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332911 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332924 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332935 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332967 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332982 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.332999 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333011 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333024 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333039 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333053 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333065 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333078 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333090 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333106 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333120 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333133 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333147 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333160 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333179 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333194 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333209 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333222 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333235 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333291 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333307 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333357 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333375 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333389 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333407 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333422 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.333439 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336610 4848 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336704 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336748 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336780 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336811 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336846 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336878 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336907 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.336938 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337011 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337039 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337069 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337101 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337130 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337172 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337206 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337236 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337268 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337303 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337336 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337364 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337397 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337427 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337455 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337482 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337510 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337537 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337565 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337593 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337620 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337647 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337674 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337704 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337758 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337788 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337819 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337853 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337884 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337914 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.337977 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338012 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338041 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338069 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338096 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338122 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338154 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338181 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338207 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338233 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338320 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338347 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338375 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338402 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338436 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338466 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338493 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338520 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338548 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338576 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338603 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338635 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338662 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338690 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338716 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338742 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338769 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338807 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338837 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338867 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338900 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.338932 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339001 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339043 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339072 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339103 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339133 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339165 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339194 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339223 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339249 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339275 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339303 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339331 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339356 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339383 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339414 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339440 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339466 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339493 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339519 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339545 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339572 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339596 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339621 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339653 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339681 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339708 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339799 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339829 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339854 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339880 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339912 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.339937 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340015 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340053 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340082 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340109 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340137 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340164 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340190 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340215 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340241 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340267 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340294 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340319 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340346 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340372 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340401 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340426 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340453 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340481 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340510 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340538 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340565 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340589 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340616 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340642 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340667 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340692 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340716 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340744 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340769 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340795 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340824 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340851 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340883 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340909 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340934 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.340999 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341042 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341074 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341100 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341126 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341152 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341183 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341208 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341231 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341259 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341287 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341318 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341344 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341385 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341413 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341442 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341469 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341495 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341523 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341554 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341583 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341609 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341637 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341664 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341726 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341758 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341783 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341816 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341846 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341875 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341901 4848 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341927 4848 reconstruct.go:97] "Volume reconstruction finished" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.341978 4848 reconciler.go:26] "Reconciler: start to sync state" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.353732 4848 manager.go:324] Recovery completed Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.368744 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.373720 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.373783 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.373795 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.376374 4848 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.376396 4848 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.376419 4848 state_mem.go:36] "Initialized new in-memory state store" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.389064 4848 policy_none.go:49] "None policy: Start" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.389063 4848 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.390374 4848 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.390410 4848 state_mem.go:35] "Initializing new in-memory state store" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.392055 4848 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.392097 4848 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.392125 4848 kubelet.go:2335] "Starting kubelet main sync loop" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.392174 4848 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.393277 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.393481 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.416847 4848 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.450481 4848 manager.go:334] "Starting Device Plugin manager" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.450536 4848 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.450550 4848 server.go:79] "Starting device plugin registration server" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.451007 4848 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.451029 4848 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.451226 4848 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.451307 4848 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.451321 4848 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.458669 4848 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.493019 4848 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.493123 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.494118 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.494174 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.494186 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.494426 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.494659 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.494701 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.495528 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.495574 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.495585 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.496276 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.496328 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.496345 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.496482 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.496656 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.496699 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497159 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497187 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497198 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497281 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497515 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497559 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497782 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497803 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.497813 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498099 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498127 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498137 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498241 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498429 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498460 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498485 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498495 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498499 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.498975 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499008 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499018 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499135 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499157 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499621 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499644 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499655 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499839 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499854 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.499863 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.521416 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="400ms" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549261 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549324 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549383 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549428 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549468 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549497 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549516 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549544 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.549898 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.550086 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.550458 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.550560 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.550591 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.550617 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.550640 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.551332 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.553141 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.553211 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.553231 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.553271 4848 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.553938 4848 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.5:6443: connect: connection refused" node="crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652198 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652266 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652302 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652333 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652369 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652400 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652430 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652460 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652528 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652529 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652634 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652625 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652693 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652693 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652557 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652640 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652792 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652760 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652537 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652802 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652722 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652883 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652845 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652938 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652976 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.652999 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.653053 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.653053 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.653191 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.653305 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.753437 4848 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.5:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e0744dc4a9e56 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 13:48:16.310861398 +0000 UTC m=+0.253357966,LastTimestamp:2025-12-04 13:48:16.310861398 +0000 UTC m=+0.253357966,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.754079 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.756171 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.756265 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.756285 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.756326 4848 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.756746 4848 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.5:6443: connect: connection refused" node="crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.838937 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.864355 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.870684 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-561689fe542fdb9e77c09f15fd5a6660d4512b242d0ad888268bc0f460ad5b18 WatchSource:0}: Error finding container 561689fe542fdb9e77c09f15fd5a6660d4512b242d0ad888268bc0f460ad5b18: Status 404 returned error can't find the container with id 561689fe542fdb9e77c09f15fd5a6660d4512b242d0ad888268bc0f460ad5b18 Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.893210 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.898626 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-65531492b97bff726e8f98771f50c2a34da599bfd7d29c87c1d9217d3d420db5 WatchSource:0}: Error finding container 65531492b97bff726e8f98771f50c2a34da599bfd7d29c87c1d9217d3d420db5: Status 404 returned error can't find the container with id 65531492b97bff726e8f98771f50c2a34da599bfd7d29c87c1d9217d3d420db5 Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.914785 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-7eff1f11f0869fb4daa269bf6c9eb79dea1ee2d237d65dad08badf1febb98353 WatchSource:0}: Error finding container 7eff1f11f0869fb4daa269bf6c9eb79dea1ee2d237d65dad08badf1febb98353: Status 404 returned error can't find the container with id 7eff1f11f0869fb4daa269bf6c9eb79dea1ee2d237d65dad08badf1febb98353 Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.915900 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: I1204 13:48:16.922288 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 13:48:16 crc kubenswrapper[4848]: E1204 13:48:16.922725 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="800ms" Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.931687 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-92ab15de33d912be26d0ab7c4b9f6dd424f3d80fcf0242eac827c54bf1906235 WatchSource:0}: Error finding container 92ab15de33d912be26d0ab7c4b9f6dd424f3d80fcf0242eac827c54bf1906235: Status 404 returned error can't find the container with id 92ab15de33d912be26d0ab7c4b9f6dd424f3d80fcf0242eac827c54bf1906235 Dec 04 13:48:16 crc kubenswrapper[4848]: W1204 13:48:16.945685 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8ac39ff31728e8edbbbb40380ed9f79406ce93a1ebdc6b8cc60a6d1edc420356 WatchSource:0}: Error finding container 8ac39ff31728e8edbbbb40380ed9f79406ce93a1ebdc6b8cc60a6d1edc420356: Status 404 returned error can't find the container with id 8ac39ff31728e8edbbbb40380ed9f79406ce93a1ebdc6b8cc60a6d1edc420356 Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.157242 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.159269 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.159349 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.159370 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.159425 4848 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:48:17 crc kubenswrapper[4848]: E1204 13:48:17.160250 4848 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.5:6443: connect: connection refused" node="crc" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.314366 4848 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:17 crc kubenswrapper[4848]: W1204 13:48:17.375492 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:17 crc kubenswrapper[4848]: E1204 13:48:17.375570 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:17 crc kubenswrapper[4848]: W1204 13:48:17.379230 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:17 crc kubenswrapper[4848]: E1204 13:48:17.379267 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.397488 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.397589 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"561689fe542fdb9e77c09f15fd5a6660d4512b242d0ad888268bc0f460ad5b18"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.397869 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.399473 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.399514 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.399525 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.400510 4848 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb" exitCode=0 Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.400589 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.400616 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8ac39ff31728e8edbbbb40380ed9f79406ce93a1ebdc6b8cc60a6d1edc420356"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.400732 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.401471 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.401498 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.401509 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.402747 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.402780 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"92ab15de33d912be26d0ab7c4b9f6dd424f3d80fcf0242eac827c54bf1906235"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.402847 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.403858 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.403987 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.404017 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.404864 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.404919 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7eff1f11f0869fb4daa269bf6c9eb79dea1ee2d237d65dad08badf1febb98353"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.406925 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.407016 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"65531492b97bff726e8f98771f50c2a34da599bfd7d29c87c1d9217d3d420db5"} Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.407153 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.408217 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.408269 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.408286 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.410230 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.411209 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.411258 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.411280 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:17 crc kubenswrapper[4848]: E1204 13:48:17.729367 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="1.6s" Dec 04 13:48:17 crc kubenswrapper[4848]: W1204 13:48:17.878843 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:17 crc kubenswrapper[4848]: E1204 13:48:17.878927 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:17 crc kubenswrapper[4848]: W1204 13:48:17.927305 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.5:6443: connect: connection refused Dec 04 13:48:17 crc kubenswrapper[4848]: E1204 13:48:17.927439 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.5:6443: connect: connection refused" logger="UnhandledError" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.962142 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.963851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.963911 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.963924 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:17 crc kubenswrapper[4848]: I1204 13:48:17.963981 4848 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:48:17 crc kubenswrapper[4848]: E1204 13:48:17.964661 4848 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.5:6443: connect: connection refused" node="crc" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.347495 4848 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.410874 4848 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661" exitCode=0 Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.411005 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.411239 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.412705 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.412764 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.412786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.415114 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.415166 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.415178 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.415262 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.416378 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.416428 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.416448 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.417967 4848 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6" exitCode=0 Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.418065 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.418126 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.418150 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.418171 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.419338 4848 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73" exitCode=0 Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.419413 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.419575 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.420507 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.420690 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.420713 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.420905 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5"} Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.421007 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.423895 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.423968 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.423986 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:18 crc kubenswrapper[4848]: I1204 13:48:18.702383 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.427077 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c"} Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.427136 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2"} Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.427149 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2"} Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.427275 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.428508 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.428549 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.428561 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.432367 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.432340 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b"} Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.432572 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5"} Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.441067 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.441142 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.441164 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.442538 4848 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a" exitCode=0 Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.442683 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a"} Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.442813 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.442903 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.444579 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.444639 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.444659 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.444676 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.444691 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.444690 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.565601 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.567666 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.567711 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.567733 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:19 crc kubenswrapper[4848]: I1204 13:48:19.567776 4848 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.397462 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.449918 4848 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7" exitCode=0 Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.450005 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7"} Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.450146 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.450186 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.450208 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.450214 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.451859 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.451900 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.451915 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.452800 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.452844 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.452861 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.452866 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.452921 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.452996 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:20 crc kubenswrapper[4848]: I1204 13:48:20.766841 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.072041 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.458848 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57"} Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.458924 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638"} Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.459002 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866"} Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.459014 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.459052 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.460443 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.460459 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.460486 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.460499 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.460502 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.460693 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.703186 4848 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 13:48:21 crc kubenswrapper[4848]: I1204 13:48:21.703401 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.469058 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795"} Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.469136 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee"} Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.469492 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.469975 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.471202 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.471278 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.471299 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.471538 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.471613 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:22 crc kubenswrapper[4848]: I1204 13:48:22.471634 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.320711 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.471723 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.471764 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.473383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.473427 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.473442 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.473383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.473486 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:23 crc kubenswrapper[4848]: I1204 13:48:23.473507 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:26 crc kubenswrapper[4848]: E1204 13:48:26.458760 4848 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 13:48:26 crc kubenswrapper[4848]: I1204 13:48:26.836014 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 04 13:48:26 crc kubenswrapper[4848]: I1204 13:48:26.836267 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:26 crc kubenswrapper[4848]: I1204 13:48:26.837793 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:26 crc kubenswrapper[4848]: I1204 13:48:26.837855 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:26 crc kubenswrapper[4848]: I1204 13:48:26.837874 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.260834 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.261127 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.262778 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.262848 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.262868 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.315206 4848 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:48:28 crc kubenswrapper[4848]: E1204 13:48:28.349453 4848 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.354083 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.361420 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.486547 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.487839 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.487892 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.487910 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.494017 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.523515 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.523806 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.525455 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.525520 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:28 crc kubenswrapper[4848]: I1204 13:48:28.525538 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.253973 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.254267 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.256042 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.256115 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.256132 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:29 crc kubenswrapper[4848]: E1204 13:48:29.329982 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.489283 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.491011 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.491078 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.491099 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:29 crc kubenswrapper[4848]: E1204 13:48:29.569725 4848 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 04 13:48:29 crc kubenswrapper[4848]: W1204 13:48:29.643937 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.644168 4848 trace.go:236] Trace[918451366]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 13:48:19.642) (total time: 10001ms): Dec 04 13:48:29 crc kubenswrapper[4848]: Trace[918451366]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:48:29.643) Dec 04 13:48:29 crc kubenswrapper[4848]: Trace[918451366]: [10.001970511s] [10.001970511s] END Dec 04 13:48:29 crc kubenswrapper[4848]: E1204 13:48:29.644229 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:48:29 crc kubenswrapper[4848]: W1204 13:48:29.846145 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:48:29 crc kubenswrapper[4848]: I1204 13:48:29.846286 4848 trace.go:236] Trace[1792772648]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 13:48:19.845) (total time: 10001ms): Dec 04 13:48:29 crc kubenswrapper[4848]: Trace[1792772648]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:48:29.846) Dec 04 13:48:29 crc kubenswrapper[4848]: Trace[1792772648]: [10.00118584s] [10.00118584s] END Dec 04 13:48:29 crc kubenswrapper[4848]: E1204 13:48:29.846336 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:48:30 crc kubenswrapper[4848]: W1204 13:48:30.066215 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:48:30 crc kubenswrapper[4848]: I1204 13:48:30.066343 4848 trace.go:236] Trace[102741584]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 13:48:20.064) (total time: 10001ms): Dec 04 13:48:30 crc kubenswrapper[4848]: Trace[102741584]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:48:30.066) Dec 04 13:48:30 crc kubenswrapper[4848]: Trace[102741584]: [10.001605026s] [10.001605026s] END Dec 04 13:48:30 crc kubenswrapper[4848]: E1204 13:48:30.066379 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:48:30 crc kubenswrapper[4848]: W1204 13:48:30.211198 4848 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 13:48:30 crc kubenswrapper[4848]: I1204 13:48:30.211335 4848 trace.go:236] Trace[1390463610]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 13:48:20.209) (total time: 10002ms): Dec 04 13:48:30 crc kubenswrapper[4848]: Trace[1390463610]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (13:48:30.211) Dec 04 13:48:30 crc kubenswrapper[4848]: Trace[1390463610]: [10.002182699s] [10.002182699s] END Dec 04 13:48:30 crc kubenswrapper[4848]: E1204 13:48:30.211375 4848 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 13:48:30 crc kubenswrapper[4848]: I1204 13:48:30.492668 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:30 crc kubenswrapper[4848]: I1204 13:48:30.494067 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:30 crc kubenswrapper[4848]: I1204 13:48:30.494113 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:30 crc kubenswrapper[4848]: I1204 13:48:30.494130 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:30 crc kubenswrapper[4848]: I1204 13:48:30.776666 4848 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 13:48:30 crc kubenswrapper[4848]: I1204 13:48:30.776736 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 13:48:31 crc kubenswrapper[4848]: I1204 13:48:31.703344 4848 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 13:48:31 crc kubenswrapper[4848]: I1204 13:48:31.703517 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 13:48:32 crc kubenswrapper[4848]: I1204 13:48:32.376369 4848 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 13:48:32 crc kubenswrapper[4848]: I1204 13:48:32.392921 4848 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 04 13:48:32 crc kubenswrapper[4848]: I1204 13:48:32.770187 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:32 crc kubenswrapper[4848]: I1204 13:48:32.771523 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:32 crc kubenswrapper[4848]: I1204 13:48:32.771598 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:32 crc kubenswrapper[4848]: I1204 13:48:32.771611 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:32 crc kubenswrapper[4848]: I1204 13:48:32.771643 4848 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:48:32 crc kubenswrapper[4848]: E1204 13:48:32.806900 4848 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 04 13:48:33 crc kubenswrapper[4848]: I1204 13:48:33.678582 4848 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 13:48:33 crc kubenswrapper[4848]: I1204 13:48:33.693524 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:33 crc kubenswrapper[4848]: I1204 13:48:33.703939 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.316054 4848 apiserver.go:52] "Watching apiserver" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.319928 4848 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.320552 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.321102 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.321253 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.321437 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.321499 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:34 crc kubenswrapper[4848]: E1204 13:48:34.321566 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:34 crc kubenswrapper[4848]: E1204 13:48:34.321593 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.322167 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.323248 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:34 crc kubenswrapper[4848]: E1204 13:48:34.323464 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.325745 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.325926 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.326185 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.326263 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.326337 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.326808 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.327010 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.327047 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.329355 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.367816 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.386191 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.407249 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.417999 4848 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.424315 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.429599 4848 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.442376 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.459224 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.474933 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.491793 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.504240 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.511532 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: E1204 13:48:34.512378 4848 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.534712 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.551220 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:34 crc kubenswrapper[4848]: I1204 13:48:34.566899 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.160727 4848 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.329181 4848 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.507173 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.780086 4848 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.828655 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.868939 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880637 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880687 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880717 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880847 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880873 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880896 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880922 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880972 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.880999 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881023 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881047 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881070 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881069 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.881097 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:48:36.381070371 +0000 UTC m=+20.323567019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881141 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881174 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881194 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881209 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881225 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881247 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881271 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881294 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881322 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881339 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881332 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881360 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881406 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881442 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881486 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881519 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881556 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881594 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881622 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881630 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881667 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881653 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881703 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881734 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881763 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881793 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881795 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881823 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881854 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.881979 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882012 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882044 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882673 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882742 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882785 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882824 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882864 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882900 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882936 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883010 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883047 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883083 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883122 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883159 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883192 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883227 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883261 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883295 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883329 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883364 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883397 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883439 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883472 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883506 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883538 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883572 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883606 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883641 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883678 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883717 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882084 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882112 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882170 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882152 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882323 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882382 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882384 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882406 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882434 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882579 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882828 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882833 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.882848 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883046 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883053 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883164 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883287 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883347 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883533 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.886127 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883553 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883604 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883673 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.883682 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.884026 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.884303 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.884391 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.884731 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.884734 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.884863 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.885072 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.885165 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.885168 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.885291 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.885494 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.885619 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.885683 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.885920 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.886300 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888037 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888088 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888126 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888165 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888206 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888244 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888296 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888333 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888370 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888409 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888447 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888449 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888486 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888523 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888559 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888621 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888663 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888704 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888744 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888813 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888850 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888886 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888924 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.888999 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889061 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889101 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889134 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889171 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889207 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889244 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889281 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889322 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889359 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889403 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889437 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889475 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889508 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889543 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889578 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889611 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889646 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889682 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889719 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889753 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889822 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889855 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889892 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889930 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.889998 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890035 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890076 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890112 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890147 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890181 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890215 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890248 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890283 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890318 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890352 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890388 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890424 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890458 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890496 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890532 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890565 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890611 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890648 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890684 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890720 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890754 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890794 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890828 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890863 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890897 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890931 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.890988 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891025 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891062 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891096 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891129 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891163 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891199 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891234 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891269 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891303 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891340 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891376 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891411 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891445 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891483 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891518 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891554 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891589 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891626 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891627 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891662 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891700 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891738 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891778 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891834 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891869 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891905 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891942 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892003 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.891992 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892041 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892082 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892119 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892156 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892193 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892230 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892266 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892303 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892340 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892376 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892415 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892454 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892489 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892525 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892563 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892599 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892642 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892681 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892721 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892757 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892793 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892828 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892865 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892902 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.893793 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.894066 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.894102 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.894418 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.894676 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.895074 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.895299 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.895503 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.895686 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.896117 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.897343 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.897718 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.898149 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.898541 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.899528 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.899730 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.892942 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.900059 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.900071 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.900227 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.900330 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.900483 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.900607 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.903694 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.903995 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904049 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904086 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904112 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904144 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904175 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904201 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904236 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904270 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904288 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904309 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904331 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904357 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904391 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904484 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904497 4848 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904508 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904519 4848 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904530 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904541 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904554 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904565 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904576 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904587 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904597 4848 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904607 4848 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904617 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904627 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904637 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904647 4848 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904657 4848 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904667 4848 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904677 4848 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904687 4848 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904696 4848 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904707 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904717 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904727 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904738 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904748 4848 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904760 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904774 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904785 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904798 4848 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904808 4848 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904818 4848 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904829 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904838 4848 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904848 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904857 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904866 4848 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904875 4848 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904886 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904898 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904907 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904917 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904927 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904937 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904963 4848 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904973 4848 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904984 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.904993 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905004 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905014 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905024 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905033 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905043 4848 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905054 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905067 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905082 4848 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905094 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905105 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905115 4848 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905126 4848 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905136 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905146 4848 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905155 4848 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905166 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905177 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905186 4848 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905195 4848 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905204 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905214 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.905223 4848 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.905766 4848 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.905835 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:36.405819168 +0000 UTC m=+20.348315696 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.906194 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.906540 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.906771 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.906855 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.906930 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.907039 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.907501 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.907569 4848 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.907601 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:36.407593741 +0000 UTC m=+20.350090269 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.907741 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.907766 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.907922 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.908003 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.907979 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.908361 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.908427 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.908453 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.907338 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.908706 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.910648 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.910710 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.912048 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.912236 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.912325 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.912345 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.912644 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.912811 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.912853 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.913010 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.913203 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.913792 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.913869 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.913777 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.914358 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.914540 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.914870 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.915694 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.915752 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916021 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916085 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916137 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916230 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916316 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916471 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916629 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916668 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916721 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.916725 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.918121 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.918416 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.918530 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.918821 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.918904 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.919095 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.919357 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.919531 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.919699 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.907052 4848 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.920019 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.920356 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.920361 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.920707 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.920888 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.921036 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.921382 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.921632 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.921650 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.922545 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.922847 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.923029 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.923179 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.923569 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.923793 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.923856 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.924138 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.924388 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.924747 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.925050 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.925654 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.926066 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.926394 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.926587 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.927102 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.927620 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.928140 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.930711 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.930829 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.931390 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.932560 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.932562 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.937683 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.938178 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.938818 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.939099 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.939121 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.939136 4848 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.939151 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.939200 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:36.439176986 +0000 UTC m=+20.381673734 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.940645 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.940668 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.940685 4848 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:35 crc kubenswrapper[4848]: E1204 13:48:35.940728 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:36.440710104 +0000 UTC m=+20.383206632 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.940882 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.941791 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.942348 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.943207 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.943357 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.943694 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.950424 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.950934 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.951183 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.953501 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.954009 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.954067 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.954410 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.954581 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.954073 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.954704 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.955840 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.955875 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957089 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957129 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957173 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957383 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957409 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957453 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957798 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957841 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.957876 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.958881 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.959649 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.960309 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.960400 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.960486 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.962590 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.962577 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.962751 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.962803 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.962963 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.963430 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.963511 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.963782 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.971788 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.974403 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.981884 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.993650 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:35 crc kubenswrapper[4848]: I1204 13:48:35.998566 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.006666 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007034 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007077 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007114 4848 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007125 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007133 4848 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007142 4848 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007151 4848 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007162 4848 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007172 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007190 4848 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007199 4848 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007207 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007215 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007224 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007231 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007239 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007246 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007254 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007262 4848 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007269 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007277 4848 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007285 4848 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007295 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007302 4848 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007310 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007317 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007325 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007334 4848 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007344 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007353 4848 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007363 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007372 4848 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007380 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007388 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007397 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007406 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007415 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007424 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007434 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007442 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007450 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007458 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007466 4848 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007475 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007485 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007494 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007502 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007512 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007521 4848 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007529 4848 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007540 4848 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007547 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007555 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007565 4848 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007573 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007582 4848 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007590 4848 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007598 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007607 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007616 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007625 4848 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007634 4848 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007643 4848 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007671 4848 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007704 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007713 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007721 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007729 4848 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007737 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007759 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007768 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007775 4848 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007784 4848 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007794 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007802 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007810 4848 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007818 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007796 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007827 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007914 4848 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007927 4848 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007939 4848 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007963 4848 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007975 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007987 4848 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.007998 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008009 4848 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008020 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008033 4848 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008043 4848 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008053 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008063 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008073 4848 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008086 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008097 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008108 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008118 4848 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008127 4848 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008137 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008146 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008156 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008166 4848 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008175 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008185 4848 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008194 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008205 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008215 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008224 4848 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008233 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008244 4848 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008259 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008273 4848 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008284 4848 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008295 4848 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008301 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008306 4848 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008347 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008369 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008379 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008388 4848 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008409 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008419 4848 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008427 4848 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008435 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008443 4848 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008451 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.008460 4848 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.015217 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.109876 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.142427 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.156328 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 13:48:36 crc kubenswrapper[4848]: W1204 13:48:36.163843 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-1368825cbffdac9cfed7a8c8bdb4584f601fdaeddb5f6debf3008dee4eaa43cb WatchSource:0}: Error finding container 1368825cbffdac9cfed7a8c8bdb4584f601fdaeddb5f6debf3008dee4eaa43cb: Status 404 returned error can't find the container with id 1368825cbffdac9cfed7a8c8bdb4584f601fdaeddb5f6debf3008dee4eaa43cb Dec 04 13:48:36 crc kubenswrapper[4848]: W1204 13:48:36.171177 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-6d23614bf3172eec1a821442195f2defe81c47d08656fb2dab1c7f13557bb809 WatchSource:0}: Error finding container 6d23614bf3172eec1a821442195f2defe81c47d08656fb2dab1c7f13557bb809: Status 404 returned error can't find the container with id 6d23614bf3172eec1a821442195f2defe81c47d08656fb2dab1c7f13557bb809 Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.172279 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 13:48:36 crc kubenswrapper[4848]: W1204 13:48:36.197373 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-1ad6b947fe607580fabd7a9f88572f4b7915f9f1b74ef9d869644f73a3ce1893 WatchSource:0}: Error finding container 1ad6b947fe607580fabd7a9f88572f4b7915f9f1b74ef9d869644f73a3ce1893: Status 404 returned error can't find the container with id 1ad6b947fe607580fabd7a9f88572f4b7915f9f1b74ef9d869644f73a3ce1893 Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.392618 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.392655 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.393020 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.392698 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.393091 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.393199 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.400440 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.401190 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.401828 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.402467 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.403932 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.404563 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.405400 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.405970 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.406633 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.407196 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.407690 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.408418 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.409025 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.409639 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.410185 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.410708 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.411302 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.411748 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.411904 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.412013 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.412092 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:48:37.412054601 +0000 UTC m=+21.354551179 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.412145 4848 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.412147 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.412198 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:37.412184874 +0000 UTC m=+21.354681402 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.412285 4848 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.412335 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:37.412327109 +0000 UTC m=+21.354823727 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.412298 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.414971 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.415915 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.419570 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.421073 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.421547 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.422677 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.423187 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.424376 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.425247 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.425800 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.427242 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.427583 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.427816 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.428941 4848 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.429099 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.431157 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.432294 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.432794 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.434604 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.435345 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.436323 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.437036 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.438246 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.438722 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.439714 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.439786 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.440798 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.441475 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.442336 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.443258 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.444240 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.445168 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.445698 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.446572 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.447103 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.447687 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.448762 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.449267 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.449744 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.462038 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.477595 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.489891 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.510995 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1ad6b947fe607580fabd7a9f88572f4b7915f9f1b74ef9d869644f73a3ce1893"} Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.512656 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.512713 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.512847 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.512883 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.512900 4848 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.512914 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.512975 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.512982 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:37.512942685 +0000 UTC m=+21.455439233 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.512995 4848 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:36 crc kubenswrapper[4848]: E1204 13:48:36.513069 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:37.513042508 +0000 UTC m=+21.455539036 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.513494 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf"} Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.513531 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf"} Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.513547 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6d23614bf3172eec1a821442195f2defe81c47d08656fb2dab1c7f13557bb809"} Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.515772 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2"} Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.516255 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1368825cbffdac9cfed7a8c8bdb4584f601fdaeddb5f6debf3008dee4eaa43cb"} Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.528074 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.540101 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.549785 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.558252 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.569132 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.577503 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.585913 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.596916 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.605818 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.623334 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.639833 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.653937 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.667828 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:36 crc kubenswrapper[4848]: I1204 13:48:36.683593 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:37 crc kubenswrapper[4848]: I1204 13:48:37.421667 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:37 crc kubenswrapper[4848]: I1204 13:48:37.421871 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.421889 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:48:39.421862914 +0000 UTC m=+23.364359442 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:48:37 crc kubenswrapper[4848]: I1204 13:48:37.421986 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.421997 4848 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.422074 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:39.422054208 +0000 UTC m=+23.364550736 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.422223 4848 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.422360 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:39.422335095 +0000 UTC m=+23.364831623 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:37 crc kubenswrapper[4848]: I1204 13:48:37.523658 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:37 crc kubenswrapper[4848]: I1204 13:48:37.523713 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.523874 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.523923 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.523884 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.523940 4848 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.523972 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.523989 4848 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.524035 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:39.524011849 +0000 UTC m=+23.466508587 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:37 crc kubenswrapper[4848]: E1204 13:48:37.524064 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:39.52405106 +0000 UTC m=+23.466547868 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.393426 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.393532 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:38 crc kubenswrapper[4848]: E1204 13:48:38.393626 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.393442 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:38 crc kubenswrapper[4848]: E1204 13:48:38.393757 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:38 crc kubenswrapper[4848]: E1204 13:48:38.393874 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.565250 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.586781 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.588881 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.592537 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.645011 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.663035 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.680906 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.695698 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.710538 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.712172 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.717901 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.721864 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.728409 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.747877 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.769552 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.788059 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.805973 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.817588 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.829394 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.844481 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.861766 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:38 crc kubenswrapper[4848]: I1204 13:48:38.878396 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:38Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.208043 4848 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.210440 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.210537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.210556 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.210624 4848 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.221415 4848 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.221806 4848 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.223429 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.223486 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.223504 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.223533 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.223552 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.252454 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.258752 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.258797 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.258820 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.258847 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.258866 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.281531 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.287511 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.287558 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.287575 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.287601 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.287619 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.309462 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.314544 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.314592 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.314610 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.314633 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.314649 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.335730 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.340518 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.340568 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.340586 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.340612 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.340630 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.361523 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.361750 4848 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.364062 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.364120 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.364146 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.364174 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.364195 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.441418 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.441589 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.441634 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:48:43.44158538 +0000 UTC m=+27.384081938 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.441704 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.441777 4848 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.441798 4848 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.441857 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:43.441835186 +0000 UTC m=+27.384331754 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.441891 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:43.441875747 +0000 UTC m=+27.384372315 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.467264 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.467344 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.467415 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.467445 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.467469 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.528414 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96"} Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.539371 4848 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.542659 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.542729 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.542890 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.542916 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.542932 4848 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.543080 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:43.543059329 +0000 UTC m=+27.485555877 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.543113 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.543328 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.543360 4848 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.543473 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:43.543434638 +0000 UTC m=+27.485931216 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:39 crc kubenswrapper[4848]: E1204 13:48:39.543558 4848 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.550413 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.566156 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.570521 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.570573 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.570587 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.570609 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.570625 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.578910 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.593438 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.608282 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.641517 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.659975 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.673189 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.673234 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.673251 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.673271 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.673286 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.677656 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.691264 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.775837 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.775902 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.775922 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.776014 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.776036 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.879268 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.879331 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.879345 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.879366 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.879381 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.983157 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.983255 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.983285 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.983323 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:39 crc kubenswrapper[4848]: I1204 13:48:39.983352 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:39Z","lastTransitionTime":"2025-12-04T13:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.086193 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.086257 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.086271 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.086295 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.086310 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.189941 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.190040 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.190062 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.190101 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.190144 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.293359 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.293399 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.293409 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.293424 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.293434 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.393394 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.393426 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:40 crc kubenswrapper[4848]: E1204 13:48:40.393546 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.393392 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:40 crc kubenswrapper[4848]: E1204 13:48:40.393702 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:40 crc kubenswrapper[4848]: E1204 13:48:40.393886 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.396259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.396414 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.396443 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.396475 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.396513 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.499353 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.499426 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.499445 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.499473 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.499491 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.602415 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.602503 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.602519 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.602546 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.602568 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.705677 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.705740 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.705753 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.705773 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.705785 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.809066 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.809117 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.809130 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.809150 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.809169 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.911604 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.911670 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.911692 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.911721 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:40 crc kubenswrapper[4848]: I1204 13:48:40.911739 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:40Z","lastTransitionTime":"2025-12-04T13:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.014856 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.014989 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.015017 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.015050 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.015073 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.119211 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.119286 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.119305 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.119335 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.119354 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.222505 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.222575 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.222593 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.222621 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.222641 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.326016 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.326058 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.326068 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.326086 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.326097 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.428333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.428425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.428450 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.428480 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.428499 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.531742 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.531786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.531797 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.531821 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.531834 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.634910 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.635016 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.635036 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.635064 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.635083 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.738274 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.738345 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.738366 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.738396 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.738418 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.842340 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.842406 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.842427 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.842450 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.842466 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.945370 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.945419 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.945432 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.945452 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.945464 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:41Z","lastTransitionTime":"2025-12-04T13:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:41 crc kubenswrapper[4848]: I1204 13:48:41.998382 4848 csr.go:261] certificate signing request csr-hsfsk is approved, waiting to be issued Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.029381 4848 csr.go:257] certificate signing request csr-hsfsk is issued Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.047906 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.047964 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.047975 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.047991 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.048002 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.156129 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.156176 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.156186 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.156202 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.156213 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.258967 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.258994 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.259002 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.259016 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.259025 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.360862 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.360911 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.360936 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.360969 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.360979 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.393264 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.393290 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:42 crc kubenswrapper[4848]: E1204 13:48:42.393371 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.393274 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:42 crc kubenswrapper[4848]: E1204 13:48:42.393502 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:42 crc kubenswrapper[4848]: E1204 13:48:42.393569 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.463500 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.463539 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.463548 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.463564 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.463574 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.565975 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.566016 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.566026 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.566040 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.566050 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.668134 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.668188 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.668202 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.668223 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.668235 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.741861 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-lpwtz"] Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.742180 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lpwtz" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.745088 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.745589 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.745711 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.757045 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.767814 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.771925 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.771984 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.771994 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.772015 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.772024 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.796763 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.815137 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.828937 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.841317 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.852624 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.863394 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.871608 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4b23ff2f-6e32-4b5c-9867-9988fa46d456-hosts-file\") pod \"node-resolver-lpwtz\" (UID: \"4b23ff2f-6e32-4b5c-9867-9988fa46d456\") " pod="openshift-dns/node-resolver-lpwtz" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.871745 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd5g4\" (UniqueName: \"kubernetes.io/projected/4b23ff2f-6e32-4b5c-9867-9988fa46d456-kube-api-access-qd5g4\") pod \"node-resolver-lpwtz\" (UID: \"4b23ff2f-6e32-4b5c-9867-9988fa46d456\") " pod="openshift-dns/node-resolver-lpwtz" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.874161 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.874204 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.874214 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.874232 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.874243 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.875216 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.888622 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.973081 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4b23ff2f-6e32-4b5c-9867-9988fa46d456-hosts-file\") pod \"node-resolver-lpwtz\" (UID: \"4b23ff2f-6e32-4b5c-9867-9988fa46d456\") " pod="openshift-dns/node-resolver-lpwtz" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.973122 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd5g4\" (UniqueName: \"kubernetes.io/projected/4b23ff2f-6e32-4b5c-9867-9988fa46d456-kube-api-access-qd5g4\") pod \"node-resolver-lpwtz\" (UID: \"4b23ff2f-6e32-4b5c-9867-9988fa46d456\") " pod="openshift-dns/node-resolver-lpwtz" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.973326 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4b23ff2f-6e32-4b5c-9867-9988fa46d456-hosts-file\") pod \"node-resolver-lpwtz\" (UID: \"4b23ff2f-6e32-4b5c-9867-9988fa46d456\") " pod="openshift-dns/node-resolver-lpwtz" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.976847 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.976889 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.976902 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.976921 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.976934 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:42Z","lastTransitionTime":"2025-12-04T13:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:42 crc kubenswrapper[4848]: I1204 13:48:42.990847 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd5g4\" (UniqueName: \"kubernetes.io/projected/4b23ff2f-6e32-4b5c-9867-9988fa46d456-kube-api-access-qd5g4\") pod \"node-resolver-lpwtz\" (UID: \"4b23ff2f-6e32-4b5c-9867-9988fa46d456\") " pod="openshift-dns/node-resolver-lpwtz" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.030975 4848 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-04 13:43:42 +0000 UTC, rotation deadline is 2026-09-02 06:13:18.110449155 +0000 UTC Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.031024 4848 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6520h24m35.079428516s for next certificate rotation Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.053262 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lpwtz" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.079087 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.079117 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.079125 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.079139 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.079147 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.113786 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-56gg2"] Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.114686 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-jspn5"] Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.115172 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.115702 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.115775 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vwdkp"] Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.116879 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-spvsd"] Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.117277 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.117343 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122233 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122862 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122886 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122520 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122545 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.123111 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122543 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122710 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122750 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.123323 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.123344 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.123399 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122808 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122808 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.122845 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.123036 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.128422 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.128528 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.128896 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.139787 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.156031 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.175913 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.183226 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.183263 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.183273 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.183290 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.183301 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.187613 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.205855 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.218333 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.231102 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.244653 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.256302 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.270025 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274701 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-systemd-units\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274732 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-ovn\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274751 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kth2k\" (UniqueName: \"kubernetes.io/projected/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-kube-api-access-kth2k\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274767 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274782 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-etc-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274796 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cnibin\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274812 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-systemd\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274834 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bedd115c-06c2-484b-88ab-4748c86b4531-proxy-tls\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.274857 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7z6f\" (UniqueName: \"kubernetes.io/projected/88de7a37-3579-4630-9aa7-64654e68a472-kube-api-access-h7z6f\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275030 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-var-lib-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275083 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-bin\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275109 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-etc-kubernetes\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275134 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-os-release\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275157 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/bedd115c-06c2-484b-88ab-4748c86b4531-rootfs\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275182 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275205 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-kubelet\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275225 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-system-cni-dir\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275247 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz2jc\" (UniqueName: \"kubernetes.io/projected/d8a59239-bebd-48bc-85d5-1b2062f90d8e-kube-api-access-nz2jc\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275311 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275351 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-env-overrides\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275399 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-hostroot\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275437 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-cni-binary-copy\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275465 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-slash\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275492 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-cnibin\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275536 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-conf-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275617 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-netd\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275647 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-k8s-cni-cncf-io\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275694 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-multus-certs\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275775 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-netns\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275801 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-log-socket\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275822 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-system-cni-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275869 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cni-binary-copy\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275926 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-kubelet\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.275981 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-node-log\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276002 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-ovn-kubernetes\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276026 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-config\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276045 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-script-lib\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276067 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-netns\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276118 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-cni-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276153 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-daemon-config\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276190 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88de7a37-3579-4630-9aa7-64654e68a472-ovn-node-metrics-cert\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276215 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-socket-dir-parent\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276238 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bedd115c-06c2-484b-88ab-4748c86b4531-mcd-auth-proxy-config\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276262 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm5tr\" (UniqueName: \"kubernetes.io/projected/bedd115c-06c2-484b-88ab-4748c86b4531-kube-api-access-qm5tr\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276293 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-cni-bin\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276323 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-os-release\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276348 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-cni-multus\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.276370 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.283262 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.284974 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.285013 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.285025 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.285043 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.285053 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.295594 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.308622 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.321041 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.344225 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.358852 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.373337 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377558 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-socket-dir-parent\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377596 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bedd115c-06c2-484b-88ab-4748c86b4531-mcd-auth-proxy-config\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377621 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm5tr\" (UniqueName: \"kubernetes.io/projected/bedd115c-06c2-484b-88ab-4748c86b4531-kube-api-access-qm5tr\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377671 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-cni-bin\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377706 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-os-release\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377737 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-cni-multus\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377744 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-cni-bin\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377768 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377788 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-socket-dir-parent\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377803 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-systemd-units\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377819 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-cni-multus\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377840 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-os-release\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377858 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-ovn\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377845 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-systemd-units\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377896 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kth2k\" (UniqueName: \"kubernetes.io/projected/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-kube-api-access-kth2k\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377934 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377898 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-ovn\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.377992 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-systemd\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378025 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-etc-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378056 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cnibin\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378083 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-systemd\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378091 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7z6f\" (UniqueName: \"kubernetes.io/projected/88de7a37-3579-4630-9aa7-64654e68a472-kube-api-access-h7z6f\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378149 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bedd115c-06c2-484b-88ab-4748c86b4531-proxy-tls\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378170 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378176 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/bedd115c-06c2-484b-88ab-4748c86b4531-rootfs\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378195 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-etc-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378205 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-var-lib-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378218 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cnibin\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378232 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-bin\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378269 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-bin\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378274 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-etc-kubernetes\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378309 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-os-release\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378340 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-var-lib-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378349 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bedd115c-06c2-484b-88ab-4748c86b4531-mcd-auth-proxy-config\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378311 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/bedd115c-06c2-484b-88ab-4748c86b4531-rootfs\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378358 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-etc-kubernetes\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378388 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378349 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378407 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-os-release\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378435 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-kubelet\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378467 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-system-cni-dir\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378501 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz2jc\" (UniqueName: \"kubernetes.io/projected/d8a59239-bebd-48bc-85d5-1b2062f90d8e-kube-api-access-nz2jc\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378528 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-var-lib-kubelet\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378533 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378567 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-env-overrides\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378616 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-hostroot\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378651 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-cni-binary-copy\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378536 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d8a59239-bebd-48bc-85d5-1b2062f90d8e-system-cni-dir\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378572 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-openvswitch\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378732 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-slash\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378693 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-hostroot\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378692 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-slash\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378789 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-cnibin\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378805 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-conf-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378834 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-netd\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378849 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-k8s-cni-cncf-io\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378864 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-multus-certs\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378881 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-system-cni-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378896 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cni-binary-copy\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378908 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-conf-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378922 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-netns\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378960 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-log-socket\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378945 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379005 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-cnibin\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.378982 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-config\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379062 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-script-lib\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379107 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-netns\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379110 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-kubelet\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379127 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-system-cni-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379139 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-node-log\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379142 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-kubelet\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379159 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-node-log\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379177 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-ovn-kubernetes\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379211 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-cni-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379257 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-netns\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379293 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-daemon-config\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379299 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-env-overrides\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379340 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88de7a37-3579-4630-9aa7-64654e68a472-ovn-node-metrics-cert\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379365 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-ovn-kubernetes\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379410 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-netd\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379455 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-k8s-cni-cncf-io\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379483 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-config\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379497 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-multus-certs\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379520 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-log-socket\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379544 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-host-run-netns\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379645 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-cni-binary-copy\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379651 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d8a59239-bebd-48bc-85d5-1b2062f90d8e-cni-binary-copy\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379733 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-cni-dir\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.379943 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-script-lib\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.380011 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-multus-daemon-config\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.383277 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88de7a37-3579-4630-9aa7-64654e68a472-ovn-node-metrics-cert\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.387877 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.387918 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.387933 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.387976 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.387995 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.390118 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bedd115c-06c2-484b-88ab-4748c86b4531-proxy-tls\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.393427 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.398266 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz2jc\" (UniqueName: \"kubernetes.io/projected/d8a59239-bebd-48bc-85d5-1b2062f90d8e-kube-api-access-nz2jc\") pod \"multus-additional-cni-plugins-56gg2\" (UID: \"d8a59239-bebd-48bc-85d5-1b2062f90d8e\") " pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.399301 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7z6f\" (UniqueName: \"kubernetes.io/projected/88de7a37-3579-4630-9aa7-64654e68a472-kube-api-access-h7z6f\") pod \"ovnkube-node-vwdkp\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.405282 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kth2k\" (UniqueName: \"kubernetes.io/projected/2b6bff84-ab72-4936-8c9f-0508ded8fdd8-kube-api-access-kth2k\") pod \"multus-spvsd\" (UID: \"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\") " pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.406901 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.408023 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm5tr\" (UniqueName: \"kubernetes.io/projected/bedd115c-06c2-484b-88ab-4748c86b4531-kube-api-access-qm5tr\") pod \"machine-config-daemon-jspn5\" (UID: \"bedd115c-06c2-484b-88ab-4748c86b4531\") " pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.419231 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.430839 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.436125 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.447066 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-56gg2" Dec 04 13:48:43 crc kubenswrapper[4848]: W1204 13:48:43.449940 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbedd115c_06c2_484b_88ab_4748c86b4531.slice/crio-f421369e4464d7f3bb9502c13405eb44ffa6592948cac26ca4feb6ad02ec3743 WatchSource:0}: Error finding container f421369e4464d7f3bb9502c13405eb44ffa6592948cac26ca4feb6ad02ec3743: Status 404 returned error can't find the container with id f421369e4464d7f3bb9502c13405eb44ffa6592948cac26ca4feb6ad02ec3743 Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.450288 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.460024 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.464712 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: W1204 13:48:43.469720 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8a59239_bebd_48bc_85d5_1b2062f90d8e.slice/crio-37f2b9a905d234043f98cb83b607dbbf42771f7cdf0334ed1b38c35ad73acef9 WatchSource:0}: Error finding container 37f2b9a905d234043f98cb83b607dbbf42771f7cdf0334ed1b38c35ad73acef9: Status 404 returned error can't find the container with id 37f2b9a905d234043f98cb83b607dbbf42771f7cdf0334ed1b38c35ad73acef9 Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.476418 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-spvsd" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.479906 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.480080 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.480120 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.480221 4848 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.480271 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:51.480253365 +0000 UTC m=+35.422749883 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.480645 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:48:51.480633074 +0000 UTC m=+35.423129602 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.480599 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.480711 4848 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.480738 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:51.480731657 +0000 UTC m=+35.423228185 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.492676 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.492727 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.492743 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.492764 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.492785 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.498583 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: W1204 13:48:43.511447 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b6bff84_ab72_4936_8c9f_0508ded8fdd8.slice/crio-e65323db01147a7411be83468cca5d88b1eeb832c07a0c03df32edc1c9d3b0b2 WatchSource:0}: Error finding container e65323db01147a7411be83468cca5d88b1eeb832c07a0c03df32edc1c9d3b0b2: Status 404 returned error can't find the container with id e65323db01147a7411be83468cca5d88b1eeb832c07a0c03df32edc1c9d3b0b2 Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.519745 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.540778 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spvsd" event={"ID":"2b6bff84-ab72-4936-8c9f-0508ded8fdd8","Type":"ContainerStarted","Data":"e65323db01147a7411be83468cca5d88b1eeb832c07a0c03df32edc1c9d3b0b2"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.542983 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"832c55b9c6aa9f5b5af6c6e9037974a5535e6a520ab26377463c2ee32070d0cd"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.545047 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lpwtz" event={"ID":"4b23ff2f-6e32-4b5c-9867-9988fa46d456","Type":"ContainerStarted","Data":"7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.545117 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lpwtz" event={"ID":"4b23ff2f-6e32-4b5c-9867-9988fa46d456","Type":"ContainerStarted","Data":"60ee913e3e3eabfb58a03f9f66278448fe482df8c82c03fa84293258ceb2f42a"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.546377 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerStarted","Data":"37f2b9a905d234043f98cb83b607dbbf42771f7cdf0334ed1b38c35ad73acef9"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.551355 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"f421369e4464d7f3bb9502c13405eb44ffa6592948cac26ca4feb6ad02ec3743"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.561712 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.581156 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.581228 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.581372 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.581388 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.581400 4848 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.581444 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:51.581431126 +0000 UTC m=+35.523927654 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.581780 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.581797 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.581829 4848 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:43 crc kubenswrapper[4848]: E1204 13:48:43.581856 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:48:51.581847446 +0000 UTC m=+35.524343984 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.588081 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.597466 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.597512 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.597535 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.597555 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.597569 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.618378 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.631311 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.646155 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.656421 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.667837 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.681380 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.693786 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.700344 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.700391 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.700402 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.700418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.700431 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.715165 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.727651 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.741647 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.757683 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.771249 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:43Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.803138 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.803209 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.803230 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.803256 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.803274 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.905567 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.905619 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.905633 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.905652 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:43 crc kubenswrapper[4848]: I1204 13:48:43.905662 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:43Z","lastTransitionTime":"2025-12-04T13:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.008390 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.008462 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.008481 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.008509 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.008529 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.113022 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.113062 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.113073 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.113101 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.113110 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.216251 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.216423 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.216448 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.216478 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.216501 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.319636 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.320053 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.320069 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.320091 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.320106 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.403831 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:44 crc kubenswrapper[4848]: E1204 13:48:44.404064 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.404600 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:44 crc kubenswrapper[4848]: E1204 13:48:44.404674 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.404902 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:44 crc kubenswrapper[4848]: E1204 13:48:44.404980 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.423216 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.423253 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.423262 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.423278 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.423287 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.541085 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.541122 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.541132 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.541147 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.541156 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.557081 4848 generic.go:334] "Generic (PLEG): container finished" podID="d8a59239-bebd-48bc-85d5-1b2062f90d8e" containerID="8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a" exitCode=0 Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.557184 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerDied","Data":"8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.559335 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.559385 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.561387 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spvsd" event={"ID":"2b6bff84-ab72-4936-8c9f-0508ded8fdd8","Type":"ContainerStarted","Data":"a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.562829 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594" exitCode=0 Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.562880 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.589008 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.607261 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.621409 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.634318 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.646376 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.646409 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.646420 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.646440 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.646452 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.661896 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.674840 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.690034 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.721493 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.734593 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.761379 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.764025 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.764061 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.764071 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.764086 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.764094 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.777556 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.795547 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.812141 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.827215 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.841144 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.853697 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.866567 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.866613 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.866627 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.866647 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.866667 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.873042 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.887498 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.901562 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.915901 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.926313 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.947382 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.956426 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.967261 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.969065 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.969108 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.969122 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.969140 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.969152 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:44Z","lastTransitionTime":"2025-12-04T13:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.981043 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:44 crc kubenswrapper[4848]: I1204 13:48:44.995084 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:44Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.013689 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.033758 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.073349 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.073663 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.074353 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.074465 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.074546 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.176931 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.176989 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.177002 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.177021 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.177034 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.279552 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.279872 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.279883 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.279898 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.279909 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.383089 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.383118 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.383126 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.383141 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.383149 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.493094 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.493123 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.493131 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.493146 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.493155 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.586661 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerStarted","Data":"3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.589133 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.589191 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.589204 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.602980 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.613053 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.613093 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.613102 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.613121 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.613129 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.619132 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.647284 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.662040 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.676047 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.697613 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.715752 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.715789 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.715801 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.715819 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.715832 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.728308 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.739994 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.753780 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.764838 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.776628 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.787895 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.804217 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.818173 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:45Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.819356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.819385 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.819397 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.819411 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.819420 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.921236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.921277 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.921294 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.921317 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:45 crc kubenswrapper[4848]: I1204 13:48:45.921335 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:45Z","lastTransitionTime":"2025-12-04T13:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.023530 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.023577 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.023589 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.023609 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.023622 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.125831 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.126103 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.126175 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.126237 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.126298 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.216768 4848 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.229295 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.229331 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.229343 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.229361 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.229374 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.333214 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.333297 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.333321 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.333354 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.333380 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.392417 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.392485 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:46 crc kubenswrapper[4848]: E1204 13:48:46.392561 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:46 crc kubenswrapper[4848]: E1204 13:48:46.392669 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.392797 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:46 crc kubenswrapper[4848]: E1204 13:48:46.393056 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.415693 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.426045 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.435890 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.435927 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.435936 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.435998 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.436011 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.440994 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.453178 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.466496 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.480365 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.494158 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.511354 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.529618 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.538772 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.538812 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.538822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.538837 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.538846 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.541152 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-8rt9n"] Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.541578 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.543380 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.543392 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.543498 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.543982 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.547939 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.562758 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.580034 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.594580 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.597149 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.597232 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.597256 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.599565 4848 generic.go:334] "Generic (PLEG): container finished" podID="d8a59239-bebd-48bc-85d5-1b2062f90d8e" containerID="3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017" exitCode=0 Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.599625 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerDied","Data":"3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.615028 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.629880 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2m7v\" (UniqueName: \"kubernetes.io/projected/599496f2-4367-4daf-9146-040bec1cd03d-kube-api-access-m2m7v\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.629982 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/599496f2-4367-4daf-9146-040bec1cd03d-host\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.630038 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/599496f2-4367-4daf-9146-040bec1cd03d-serviceca\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.630410 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.642151 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.642204 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.642214 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.642236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.642248 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.673443 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.689808 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.701539 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.713034 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.723177 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.731286 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2m7v\" (UniqueName: \"kubernetes.io/projected/599496f2-4367-4daf-9146-040bec1cd03d-kube-api-access-m2m7v\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.731407 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/599496f2-4367-4daf-9146-040bec1cd03d-host\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.731455 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/599496f2-4367-4daf-9146-040bec1cd03d-serviceca\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.732783 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/599496f2-4367-4daf-9146-040bec1cd03d-host\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.734097 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/599496f2-4367-4daf-9146-040bec1cd03d-serviceca\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.742397 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.745047 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.745095 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.745110 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.745132 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.745147 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.768648 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2m7v\" (UniqueName: \"kubernetes.io/projected/599496f2-4367-4daf-9146-040bec1cd03d-kube-api-access-m2m7v\") pod \"node-ca-8rt9n\" (UID: \"599496f2-4367-4daf-9146-040bec1cd03d\") " pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.774115 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.796984 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.821651 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.835442 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.847962 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.847936 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.848016 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.848029 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.848053 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.848070 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.860471 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8rt9n" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.860480 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: W1204 13:48:46.873247 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod599496f2_4367_4daf_9146_040bec1cd03d.slice/crio-0b087bffbfdc756b845474eacdda3e10b1af48b5f71ff39b143355329d8ec14a WatchSource:0}: Error finding container 0b087bffbfdc756b845474eacdda3e10b1af48b5f71ff39b143355329d8ec14a: Status 404 returned error can't find the container with id 0b087bffbfdc756b845474eacdda3e10b1af48b5f71ff39b143355329d8ec14a Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.877003 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.895851 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.951974 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.952014 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.952026 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.952040 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:46 crc kubenswrapper[4848]: I1204 13:48:46.952050 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:46Z","lastTransitionTime":"2025-12-04T13:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.054942 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.054998 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.055006 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.055020 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.055031 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.157334 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.157391 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.157412 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.157436 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.157454 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.260279 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.260323 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.260333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.260349 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.260360 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.362588 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.362623 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.362632 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.362646 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.362657 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.465369 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.465436 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.465449 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.465467 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.465481 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.568656 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.569034 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.569164 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.569295 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.569459 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.606439 4848 generic.go:334] "Generic (PLEG): container finished" podID="d8a59239-bebd-48bc-85d5-1b2062f90d8e" containerID="38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e" exitCode=0 Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.606537 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerDied","Data":"38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.608371 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8rt9n" event={"ID":"599496f2-4367-4daf-9146-040bec1cd03d","Type":"ContainerStarted","Data":"5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.608483 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8rt9n" event={"ID":"599496f2-4367-4daf-9146-040bec1cd03d","Type":"ContainerStarted","Data":"0b087bffbfdc756b845474eacdda3e10b1af48b5f71ff39b143355329d8ec14a"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.629779 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.650110 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.672745 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.672794 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.672803 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.672819 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.672829 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.697353 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.715450 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.728140 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.748592 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.766666 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.775181 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.775221 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.775230 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.775244 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.775254 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.795715 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.814520 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.828027 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.849545 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.863032 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.874552 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.877096 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.877134 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.877146 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.877167 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.877179 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.885758 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.896577 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.909874 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.923346 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.935775 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.955798 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.968795 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.979643 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.979686 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.979700 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.979718 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.979730 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:47Z","lastTransitionTime":"2025-12-04T13:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:47 crc kubenswrapper[4848]: I1204 13:48:47.992429 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.008902 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.024637 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.040073 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.054146 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.066041 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.082276 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.082324 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.082337 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.082356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.082370 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.088389 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.105175 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.119259 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.137846 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.184478 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.184526 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.184538 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.184557 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.184570 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.287036 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.287082 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.287094 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.287114 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.287126 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.389872 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.389919 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.389929 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.389964 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.389976 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.392479 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.392694 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.392909 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:48 crc kubenswrapper[4848]: E1204 13:48:48.393057 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:48 crc kubenswrapper[4848]: E1204 13:48:48.393149 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:48 crc kubenswrapper[4848]: E1204 13:48:48.392888 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.492281 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.492339 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.492356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.492381 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.492399 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.595482 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.595541 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.595558 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.595585 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.595604 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.617107 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.620663 4848 generic.go:334] "Generic (PLEG): container finished" podID="d8a59239-bebd-48bc-85d5-1b2062f90d8e" containerID="bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc" exitCode=0 Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.620714 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerDied","Data":"bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.637647 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.655092 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.673110 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.687161 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.698236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.698294 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.698309 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.698326 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.698340 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.709750 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.724087 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.736510 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.749672 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.760685 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.775689 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.786252 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.801235 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.801346 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.801377 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.801388 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.801407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.801980 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.819026 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.832535 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.844450 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:48Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.904785 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.905047 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.905113 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.905174 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:48 crc kubenswrapper[4848]: I1204 13:48:48.905229 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:48Z","lastTransitionTime":"2025-12-04T13:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.007896 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.008236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.008424 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.008604 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.008823 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.111778 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.111843 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.111864 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.111917 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.111939 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.215762 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.216107 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.216124 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.216142 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.216154 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.318242 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.318305 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.318322 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.318346 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.318364 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.416365 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.416419 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.416436 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.416456 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.416470 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: E1204 13:48:49.436393 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.441344 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.441389 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.441403 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.441423 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.441438 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: E1204 13:48:49.459327 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.464201 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.464236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.464244 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.464258 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.464267 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: E1204 13:48:49.476136 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.480141 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.480216 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.480231 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.480252 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.480267 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: E1204 13:48:49.496385 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.504130 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.504178 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.504193 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.504213 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.504227 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: E1204 13:48:49.536346 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: E1204 13:48:49.536504 4848 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.538786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.538822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.538832 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.538849 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.538862 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.628488 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerStarted","Data":"480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.641192 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.641253 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.641264 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.641282 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.641295 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.647844 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.660104 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.672650 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.683472 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.703165 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.715498 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.732985 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.743876 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.743912 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.743922 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.743937 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.743964 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.751613 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.764746 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.779118 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.789553 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.804766 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.822758 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.841143 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.846279 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.846333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.846356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.846386 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.846408 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.855255 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:49Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.950278 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.950333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.950344 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.950364 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:49 crc kubenswrapper[4848]: I1204 13:48:49.950379 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:49Z","lastTransitionTime":"2025-12-04T13:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.056572 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.056620 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.056636 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.056653 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.056663 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.159993 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.160040 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.160049 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.160065 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.160074 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.262691 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.262749 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.262766 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.262791 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.262810 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.365654 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.365702 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.365713 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.365730 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.365742 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.392657 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.392712 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.392709 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:50 crc kubenswrapper[4848]: E1204 13:48:50.392834 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:50 crc kubenswrapper[4848]: E1204 13:48:50.392922 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:50 crc kubenswrapper[4848]: E1204 13:48:50.393118 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.469676 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.469751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.469775 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.469807 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.469831 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.572889 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.572986 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.573007 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.573033 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.573051 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.638598 4848 generic.go:334] "Generic (PLEG): container finished" podID="d8a59239-bebd-48bc-85d5-1b2062f90d8e" containerID="480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9" exitCode=0 Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.638678 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerDied","Data":"480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.656344 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.675303 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.675356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.675368 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.675389 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.675405 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.679391 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.708276 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.727357 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.741722 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.761252 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.778592 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.778581 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.778655 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.778674 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.778698 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.778717 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.805143 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.832113 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.847071 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.874743 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.881071 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.881095 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.881105 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.881119 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.881128 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.892213 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.916830 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.932668 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.944309 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:50Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.983138 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.983167 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.983175 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.983191 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:50 crc kubenswrapper[4848]: I1204 13:48:50.983200 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:50Z","lastTransitionTime":"2025-12-04T13:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.085569 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.085696 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.085774 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.085843 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.085916 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.189273 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.189325 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.189337 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.189367 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.189378 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.291937 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.292008 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.292020 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.292036 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.292046 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.393883 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.394244 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.394259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.394281 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.394293 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.480939 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.481083 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.481130 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.481266 4848 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.481298 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:49:07.481259933 +0000 UTC m=+51.423756481 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.481300 4848 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.481359 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:49:07.481345695 +0000 UTC m=+51.423842233 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.481432 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:49:07.481394167 +0000 UTC m=+51.423890925 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.500737 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.500806 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.500826 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.500854 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.500872 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.582302 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.582391 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.582610 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.582670 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.582694 4848 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.582633 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.582780 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:49:07.582752173 +0000 UTC m=+51.525248761 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.582804 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.582832 4848 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:51 crc kubenswrapper[4848]: E1204 13:48:51.582915 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:49:07.582888236 +0000 UTC m=+51.525384794 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.603935 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.604030 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.604052 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.604082 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.604102 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.653214 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.653667 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.659537 4848 generic.go:334] "Generic (PLEG): container finished" podID="d8a59239-bebd-48bc-85d5-1b2062f90d8e" containerID="a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2" exitCode=0 Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.659606 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerDied","Data":"a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.676549 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.687492 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.696075 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.707066 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.707140 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.707162 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.707193 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.707216 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.733453 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.749619 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.769037 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.788459 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.810427 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.810489 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.810507 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.810535 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.810554 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.817724 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.839992 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.862404 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.883534 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.911649 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.913393 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.913433 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.913442 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.913457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.913476 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:51Z","lastTransitionTime":"2025-12-04T13:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.941361 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.955908 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.970157 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:51 crc kubenswrapper[4848]: I1204 13:48:51.984605 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.006917 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.017041 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.017082 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.017092 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.017111 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.017121 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.030548 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.052712 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.073871 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.094613 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.114575 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.123765 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.123843 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.123874 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.123908 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.123929 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.135924 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.158641 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.173201 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.188522 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.204382 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.223215 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.227626 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.227692 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.227709 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.227739 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.227760 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.250462 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.266457 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.281915 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.330672 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.330720 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.330737 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.330761 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.330778 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.392679 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.392822 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.392883 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:52 crc kubenswrapper[4848]: E1204 13:48:52.393052 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:52 crc kubenswrapper[4848]: E1204 13:48:52.393223 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:52 crc kubenswrapper[4848]: E1204 13:48:52.393431 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.433205 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.433252 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.433266 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.433285 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.433297 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.536642 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.536695 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.536713 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.536736 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.536754 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.639379 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.639430 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.639447 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.639472 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.639490 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.668176 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.668844 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" event={"ID":"d8a59239-bebd-48bc-85d5-1b2062f90d8e","Type":"ContainerStarted","Data":"7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.669131 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.689731 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.704676 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.727573 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.742557 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.742643 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.742661 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.742690 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.742709 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.763386 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.777676 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.792303 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.823639 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.836137 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.845306 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.845343 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.845354 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.845369 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.845380 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.859366 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.875369 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.891279 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.913448 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.925007 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.936486 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.947479 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.947518 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.947528 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.947545 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.947558 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:52Z","lastTransitionTime":"2025-12-04T13:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.949726 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.960776 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:52 crc kubenswrapper[4848]: I1204 13:48:52.989283 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:52Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.002734 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.014855 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.028460 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.044908 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.049351 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.049409 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.049421 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.049437 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.049447 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.057418 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.069924 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.096319 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.110769 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.129194 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.150140 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.151722 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.151768 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.151777 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.151793 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.151822 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.164208 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.179674 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.195194 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.211428 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:53Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.253793 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.253833 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.253844 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.253877 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.253893 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.356693 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.356758 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.356780 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.356807 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.356828 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.459197 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.459254 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.459267 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.459286 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.459300 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.562047 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.562090 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.562100 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.562116 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.562169 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.665683 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.665743 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.665761 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.665789 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.665807 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.672127 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.768895 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.768985 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.769006 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.769031 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.769057 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.872598 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.872655 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.872671 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.872698 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.872715 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.976373 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.976439 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.976457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.976484 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:53 crc kubenswrapper[4848]: I1204 13:48:53.976504 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:53Z","lastTransitionTime":"2025-12-04T13:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.085465 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.085724 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.085733 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.085751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.085760 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.189620 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.189675 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.189695 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.189724 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.189743 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.292682 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.292722 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.292738 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.292761 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.292776 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.393152 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.393260 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:54 crc kubenswrapper[4848]: E1204 13:48:54.393423 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.393513 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:54 crc kubenswrapper[4848]: E1204 13:48:54.393653 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:54 crc kubenswrapper[4848]: E1204 13:48:54.393828 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.396310 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.396364 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.396378 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.396401 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.396423 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.499795 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.499862 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.499875 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.499897 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.499910 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.603211 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.603283 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.603303 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.603335 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.603353 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.679249 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/0.log" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.683867 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99" exitCode=1 Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.683927 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.685278 4848 scope.go:117] "RemoveContainer" containerID="4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.705774 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.705845 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.705863 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.705892 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.705910 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.711110 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.732800 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.772229 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.790080 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.808408 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.808454 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.808468 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.808489 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.808505 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.808896 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.827052 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.841866 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.862504 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.878594 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.901199 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.911554 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.911623 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.911636 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.911659 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.911675 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:54Z","lastTransitionTime":"2025-12-04T13:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.922398 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.934127 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.954181 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.969046 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:54 crc kubenswrapper[4848]: I1204 13:48:54.992152 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:54Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:54.194710 6114 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 13:48:54.194744 6114 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:48:54.194749 6114 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:48:54.194766 6114 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:48:54.194770 6114 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:48:54.194781 6114 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:54.194815 6114 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:48:54.194808 6114 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:48:54.194824 6114 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:48:54.194838 6114 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:48:54.194839 6114 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:54.194849 6114 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:54.194862 6114 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:54.194864 6114 factory.go:656] Stopping watch factory\\\\nI1204 13:48:54.194884 6114 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:54Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.014552 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.014580 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.014593 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.014610 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.014619 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.117084 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.117143 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.117159 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.117182 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.117199 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.219701 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.219740 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.219748 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.219763 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.219774 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.287322 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7"] Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.287942 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.289924 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.290316 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.305161 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.319050 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.322177 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.322241 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.322266 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.322290 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.322306 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.339645 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.358336 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.369796 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.381061 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.390896 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.404636 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.418210 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.425022 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.425073 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.425086 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.425103 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.425115 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.431978 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/937d7bdf-0181-4852-8d9e-1eb0a0540577-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.432086 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/937d7bdf-0181-4852-8d9e-1eb0a0540577-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.432108 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg7cd\" (UniqueName: \"kubernetes.io/projected/937d7bdf-0181-4852-8d9e-1eb0a0540577-kube-api-access-kg7cd\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.432143 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/937d7bdf-0181-4852-8d9e-1eb0a0540577-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.433540 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.451236 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.468503 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.490869 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.516351 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:54Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:54.194710 6114 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 13:48:54.194744 6114 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:48:54.194749 6114 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:48:54.194766 6114 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:48:54.194770 6114 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:48:54.194781 6114 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:54.194815 6114 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:48:54.194808 6114 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:48:54.194824 6114 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:48:54.194838 6114 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:48:54.194839 6114 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:54.194849 6114 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:54.194862 6114 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:54.194864 6114 factory.go:656] Stopping watch factory\\\\nI1204 13:48:54.194884 6114 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.528007 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.528062 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.528073 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.528094 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.528105 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.531487 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.533087 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/937d7bdf-0181-4852-8d9e-1eb0a0540577-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.533125 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg7cd\" (UniqueName: \"kubernetes.io/projected/937d7bdf-0181-4852-8d9e-1eb0a0540577-kube-api-access-kg7cd\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.533159 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/937d7bdf-0181-4852-8d9e-1eb0a0540577-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.533191 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/937d7bdf-0181-4852-8d9e-1eb0a0540577-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.533840 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/937d7bdf-0181-4852-8d9e-1eb0a0540577-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.534298 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/937d7bdf-0181-4852-8d9e-1eb0a0540577-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.540879 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/937d7bdf-0181-4852-8d9e-1eb0a0540577-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.547045 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:55Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.556195 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg7cd\" (UniqueName: \"kubernetes.io/projected/937d7bdf-0181-4852-8d9e-1eb0a0540577-kube-api-access-kg7cd\") pod \"ovnkube-control-plane-749d76644c-fd6s7\" (UID: \"937d7bdf-0181-4852-8d9e-1eb0a0540577\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.608385 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" Dec 04 13:48:55 crc kubenswrapper[4848]: W1204 13:48:55.622562 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod937d7bdf_0181_4852_8d9e_1eb0a0540577.slice/crio-b92553e740b199587de627af2e0aee51465b15080835babe91fd22c7abd32db2 WatchSource:0}: Error finding container b92553e740b199587de627af2e0aee51465b15080835babe91fd22c7abd32db2: Status 404 returned error can't find the container with id b92553e740b199587de627af2e0aee51465b15080835babe91fd22c7abd32db2 Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.630434 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.630494 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.630511 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.630536 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.630554 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.687592 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" event={"ID":"937d7bdf-0181-4852-8d9e-1eb0a0540577","Type":"ContainerStarted","Data":"b92553e740b199587de627af2e0aee51465b15080835babe91fd22c7abd32db2"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.690373 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/0.log" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.693119 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.733260 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.733302 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.733314 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.733331 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.733354 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.835705 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.836110 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.836121 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.836139 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.836147 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.938444 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.938479 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.938487 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.938501 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:55 crc kubenswrapper[4848]: I1204 13:48:55.938509 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:55Z","lastTransitionTime":"2025-12-04T13:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.041713 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.041802 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.041831 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.041864 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.041886 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.144437 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.144502 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.144521 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.144553 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.144574 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.247404 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.247469 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.247501 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.247527 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.247545 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.350763 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.350821 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.350839 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.350864 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.350886 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.393291 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.393369 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.393291 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:56 crc kubenswrapper[4848]: E1204 13:48:56.393531 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:56 crc kubenswrapper[4848]: E1204 13:48:56.393624 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:56 crc kubenswrapper[4848]: E1204 13:48:56.393723 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.422277 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.439416 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.453985 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.454029 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.454044 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.454064 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.454077 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.463039 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.482587 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.493299 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.506753 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.519585 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.531067 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.551657 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.556713 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.556745 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.556786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.556804 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.556826 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.565121 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.575178 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.588051 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.600719 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.619686 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:54Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:54.194710 6114 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 13:48:54.194744 6114 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:48:54.194749 6114 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:48:54.194766 6114 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:48:54.194770 6114 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:48:54.194781 6114 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:54.194815 6114 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:48:54.194808 6114 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:48:54.194824 6114 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:48:54.194838 6114 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:48:54.194839 6114 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:54.194849 6114 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:54.194862 6114 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:54.194864 6114 factory.go:656] Stopping watch factory\\\\nI1204 13:48:54.194884 6114 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.635338 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.650586 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.659528 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.659564 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.659572 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.659588 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.659598 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.762735 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.762818 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.762831 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.762851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.762863 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.776526 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-v4k6x"] Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.777523 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:56 crc kubenswrapper[4848]: E1204 13:48:56.777782 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.799320 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.820791 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.839109 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.847119 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.847207 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d725d\" (UniqueName: \"kubernetes.io/projected/5577aeec-1fac-4adc-a566-b0110dd2477b-kube-api-access-d725d\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.856131 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.865379 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.865417 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.865426 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.865441 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.865451 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.876925 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.905418 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:54Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:54.194710 6114 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 13:48:54.194744 6114 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:48:54.194749 6114 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:48:54.194766 6114 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:48:54.194770 6114 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:48:54.194781 6114 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:54.194815 6114 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:48:54.194808 6114 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:48:54.194824 6114 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:48:54.194838 6114 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:48:54.194839 6114 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:54.194849 6114 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:54.194862 6114 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:54.194864 6114 factory.go:656] Stopping watch factory\\\\nI1204 13:48:54.194884 6114 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.918423 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.931601 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.948194 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d725d\" (UniqueName: \"kubernetes.io/projected/5577aeec-1fac-4adc-a566-b0110dd2477b-kube-api-access-d725d\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.948794 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:56 crc kubenswrapper[4848]: E1204 13:48:56.949048 4848 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:48:56 crc kubenswrapper[4848]: E1204 13:48:56.949131 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs podName:5577aeec-1fac-4adc-a566-b0110dd2477b nodeName:}" failed. No retries permitted until 2025-12-04 13:48:57.449101654 +0000 UTC m=+41.391598192 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs") pod "network-metrics-daemon-v4k6x" (UID: "5577aeec-1fac-4adc-a566-b0110dd2477b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.953114 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.967628 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.967663 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.967674 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.967691 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.967701 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:56Z","lastTransitionTime":"2025-12-04T13:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.969475 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.979414 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d725d\" (UniqueName: \"kubernetes.io/projected/5577aeec-1fac-4adc-a566-b0110dd2477b-kube-api-access-d725d\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.982210 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:56 crc kubenswrapper[4848]: I1204 13:48:56.999602 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.014340 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.036379 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.057721 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.070341 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.070399 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.070410 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.070425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.070433 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.073965 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.085634 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.172923 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.172982 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.172993 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.173010 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.173021 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.275059 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.275089 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.275098 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.275112 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.275121 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.377259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.377322 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.377335 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.377354 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.377367 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.455290 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:57 crc kubenswrapper[4848]: E1204 13:48:57.455471 4848 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:48:57 crc kubenswrapper[4848]: E1204 13:48:57.455532 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs podName:5577aeec-1fac-4adc-a566-b0110dd2477b nodeName:}" failed. No retries permitted until 2025-12-04 13:48:58.455513062 +0000 UTC m=+42.398009590 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs") pod "network-metrics-daemon-v4k6x" (UID: "5577aeec-1fac-4adc-a566-b0110dd2477b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.480109 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.480151 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.480164 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.480183 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.480195 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.581996 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.582031 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.582040 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.582053 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.582061 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.684236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.684311 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.684336 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.684366 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.684387 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.703977 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" event={"ID":"937d7bdf-0181-4852-8d9e-1eb0a0540577","Type":"ContainerStarted","Data":"6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.704082 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" event={"ID":"937d7bdf-0181-4852-8d9e-1eb0a0540577","Type":"ContainerStarted","Data":"4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.706595 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/1.log" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.707225 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/0.log" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.711151 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b" exitCode=1 Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.711215 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.711287 4848 scope.go:117] "RemoveContainer" containerID="4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.712531 4848 scope.go:117] "RemoveContainer" containerID="3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b" Dec 04 13:48:57 crc kubenswrapper[4848]: E1204 13:48:57.712914 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.723245 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.738137 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.758501 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:54Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:54.194710 6114 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 13:48:54.194744 6114 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:48:54.194749 6114 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:48:54.194766 6114 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:48:54.194770 6114 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:48:54.194781 6114 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:54.194815 6114 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:48:54.194808 6114 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:48:54.194824 6114 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:48:54.194838 6114 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:48:54.194839 6114 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:54.194849 6114 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:54.194862 6114 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:54.194864 6114 factory.go:656] Stopping watch factory\\\\nI1204 13:48:54.194884 6114 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.772329 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.787567 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.787603 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.787613 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.787628 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.787639 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.799904 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.831279 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.843612 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.866656 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.878707 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.889980 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.890026 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.890048 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.890067 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.890081 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.891985 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.902795 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.916628 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.936924 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.949893 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.964963 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.976883 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.991836 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.991870 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.991879 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.991894 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.991903 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:57Z","lastTransitionTime":"2025-12-04T13:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:57 crc kubenswrapper[4848]: I1204 13:48:57.993280 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:57Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.006178 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.020114 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.032199 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.043511 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.055291 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.067093 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.077284 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.094546 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.094595 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.094607 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.094623 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.094633 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.098389 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.114258 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.131052 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.143717 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.156539 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.172878 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.184142 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.197124 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.197170 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.197188 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.197213 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.197230 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.203342 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.224075 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e1884b522a92a96f3cb21fff6aa806454c44aa20c5eec535aabbf6c41fb0c99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:54Z\\\",\\\"message\\\":\\\"ice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:54.194710 6114 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 13:48:54.194744 6114 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 13:48:54.194749 6114 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 13:48:54.194766 6114 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 13:48:54.194770 6114 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 13:48:54.194781 6114 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:54.194815 6114 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 13:48:54.194808 6114 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 13:48:54.194824 6114 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 13:48:54.194838 6114 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 13:48:54.194839 6114 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:54.194849 6114 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:54.194862 6114 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:54.194864 6114 factory.go:656] Stopping watch factory\\\\nI1204 13:48:54.194884 6114 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:140\\\\nI1204 13:48:57.485444 6286 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:57.485459 6286 factory.go:656] Stopping watch factory\\\\nI1204 13:48:57.485470 6286 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:57.485478 6286 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:57.485484 6286 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:57.485529 6286 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:57.485843 6286 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486419 6286 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486852 6286 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486974 6286 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.487034 6286 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.239286 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.300281 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.300345 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.300362 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.300387 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.300406 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.392824 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.392884 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.392908 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.392909 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:48:58 crc kubenswrapper[4848]: E1204 13:48:58.393099 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:48:58 crc kubenswrapper[4848]: E1204 13:48:58.393258 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:48:58 crc kubenswrapper[4848]: E1204 13:48:58.393397 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:48:58 crc kubenswrapper[4848]: E1204 13:48:58.393550 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.403695 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.403751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.403764 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.403782 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.403795 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.463650 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:48:58 crc kubenswrapper[4848]: E1204 13:48:58.463910 4848 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:48:58 crc kubenswrapper[4848]: E1204 13:48:58.464077 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs podName:5577aeec-1fac-4adc-a566-b0110dd2477b nodeName:}" failed. No retries permitted until 2025-12-04 13:49:00.464043373 +0000 UTC m=+44.406539941 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs") pod "network-metrics-daemon-v4k6x" (UID: "5577aeec-1fac-4adc-a566-b0110dd2477b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.507065 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.507131 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.507170 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.507202 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.507224 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.611068 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.611121 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.611143 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.611169 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.611187 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.714581 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.714625 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.714637 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.714657 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.714669 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.718112 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/1.log" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.720080 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.723281 4848 scope.go:117] "RemoveContainer" containerID="3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b" Dec 04 13:48:58 crc kubenswrapper[4848]: E1204 13:48:58.723553 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.757983 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.773600 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.788670 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.806845 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.817588 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.817660 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.817685 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.817716 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.817740 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.823846 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.840172 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.857406 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.874383 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.888614 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.910450 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.920727 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.920798 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.920815 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.920839 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.920856 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:58Z","lastTransitionTime":"2025-12-04T13:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.932602 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.945257 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.966442 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:58 crc kubenswrapper[4848]: I1204 13:48:58.980224 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:58Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.004736 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:140\\\\nI1204 13:48:57.485444 6286 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:57.485459 6286 factory.go:656] Stopping watch factory\\\\nI1204 13:48:57.485470 6286 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:57.485478 6286 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:57.485484 6286 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:57.485529 6286 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:57.485843 6286 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486419 6286 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486852 6286 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486974 6286 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.487034 6286 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.020780 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.023813 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.023865 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.023876 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.023894 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.023905 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.037327 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.127537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.127874 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.128109 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.128294 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.128452 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.232370 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.232439 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.232456 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.232481 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.232498 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.336445 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.336506 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.336523 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.336547 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.336564 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.439748 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.439808 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.439824 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.439849 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.439868 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.543128 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.543208 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.543227 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.543253 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.543271 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.603439 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.603511 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.603536 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.603566 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.603591 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: E1204 13:48:59.620943 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.626368 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.626488 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.626502 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.626521 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.626532 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: E1204 13:48:59.644794 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.649917 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.650275 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.650453 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.650629 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.650782 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: E1204 13:48:59.672535 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.677851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.677903 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.677914 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.677934 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.677978 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: E1204 13:48:59.692711 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.697176 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.697208 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.697219 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.697236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.697249 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: E1204 13:48:59.711916 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:48:59Z is after 2025-08-24T17:21:41Z" Dec 04 13:48:59 crc kubenswrapper[4848]: E1204 13:48:59.712177 4848 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.714803 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.714846 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.714859 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.714878 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.714890 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.726880 4848 scope.go:117] "RemoveContainer" containerID="3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b" Dec 04 13:48:59 crc kubenswrapper[4848]: E1204 13:48:59.727188 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.818332 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.818417 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.818440 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.818469 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.818502 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.938669 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.938707 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.938719 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.938735 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:48:59 crc kubenswrapper[4848]: I1204 13:48:59.938747 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:48:59Z","lastTransitionTime":"2025-12-04T13:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.042388 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.042434 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.042451 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.042474 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.042493 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.145167 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.145210 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.145221 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.145239 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.145251 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.253496 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.253567 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.253588 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.253616 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.253634 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.356829 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.356900 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.356916 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.356941 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.356995 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.393081 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.393165 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:00 crc kubenswrapper[4848]: E1204 13:49:00.393219 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.393250 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.393272 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:00 crc kubenswrapper[4848]: E1204 13:49:00.393450 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:00 crc kubenswrapper[4848]: E1204 13:49:00.393629 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:00 crc kubenswrapper[4848]: E1204 13:49:00.393720 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.461357 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.461429 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.461451 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.461479 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.461499 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.488896 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:00 crc kubenswrapper[4848]: E1204 13:49:00.489159 4848 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:49:00 crc kubenswrapper[4848]: E1204 13:49:00.489246 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs podName:5577aeec-1fac-4adc-a566-b0110dd2477b nodeName:}" failed. No retries permitted until 2025-12-04 13:49:04.489222225 +0000 UTC m=+48.431718793 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs") pod "network-metrics-daemon-v4k6x" (UID: "5577aeec-1fac-4adc-a566-b0110dd2477b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.565786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.566211 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.566473 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.566655 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.566816 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.670876 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.671340 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.671590 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.671779 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.672005 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.775116 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.775625 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.775910 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.776216 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.776511 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.879663 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.879738 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.879762 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.879795 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.879817 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.983440 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.983499 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.983516 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.983541 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:00 crc kubenswrapper[4848]: I1204 13:49:00.983560 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:00Z","lastTransitionTime":"2025-12-04T13:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.087008 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.087235 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.087399 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.087504 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.087621 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.191203 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.191261 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.191280 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.191312 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.191354 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.294697 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.294757 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.294785 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.294818 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.294841 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.397619 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.398048 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.398184 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.398347 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.398468 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.501546 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.501911 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.502092 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.502233 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.502401 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.605203 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.605576 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.605771 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.606018 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.606228 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.708448 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.708505 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.708521 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.708542 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.708559 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.811856 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.811922 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.811939 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.812000 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.812022 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.914757 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.914823 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.914841 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.914867 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:01 crc kubenswrapper[4848]: I1204 13:49:01.914885 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:01Z","lastTransitionTime":"2025-12-04T13:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.017582 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.017662 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.017709 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.017739 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.017764 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.121839 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.121927 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.121990 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.122024 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.122048 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.225055 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.225318 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.225337 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.225369 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.225389 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.328599 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.329030 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.329164 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.329299 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.329418 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.393370 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:02 crc kubenswrapper[4848]: E1204 13:49:02.393543 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.393569 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.393580 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.393612 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:02 crc kubenswrapper[4848]: E1204 13:49:02.393664 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:02 crc kubenswrapper[4848]: E1204 13:49:02.393728 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:02 crc kubenswrapper[4848]: E1204 13:49:02.393772 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.432859 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.432943 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.432994 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.433022 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.433042 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.537422 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.537519 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.537550 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.537576 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.537595 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.639751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.639847 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.639869 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.639895 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.639914 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.741794 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.741845 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.741889 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.741905 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.741916 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.845528 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.845611 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.845636 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.845664 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.845683 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.948400 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.948459 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.948478 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.948508 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:02 crc kubenswrapper[4848]: I1204 13:49:02.948531 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:02Z","lastTransitionTime":"2025-12-04T13:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.052174 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.052245 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.052267 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.052298 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.052324 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.155840 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.156225 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.156383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.156522 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.156641 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.259730 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.259788 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.259803 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.259830 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.259847 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.363117 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.363189 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.363210 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.363235 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.363253 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.466927 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.467028 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.467052 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.467087 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.467111 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.570247 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.570296 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.570308 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.570325 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.570336 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.672751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.672792 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.672800 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.672815 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.672825 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.775370 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.775422 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.775435 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.775461 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.775474 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.879531 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.879584 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.879602 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.879633 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.879653 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.983054 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.983109 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.983122 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.983174 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:03 crc kubenswrapper[4848]: I1204 13:49:03.983190 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:03Z","lastTransitionTime":"2025-12-04T13:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.086540 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.086592 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.086604 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.086620 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.086633 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.190123 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.190185 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.190198 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.190219 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.190232 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.293502 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.293583 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.293600 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.293628 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.293646 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.392891 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.392995 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.393096 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:04 crc kubenswrapper[4848]: E1204 13:49:04.393294 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.393324 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:04 crc kubenswrapper[4848]: E1204 13:49:04.393494 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:04 crc kubenswrapper[4848]: E1204 13:49:04.393610 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:04 crc kubenswrapper[4848]: E1204 13:49:04.393681 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.396415 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.396448 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.396457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.396474 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.396489 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.500407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.500490 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.500509 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.500537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.500554 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.538298 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:04 crc kubenswrapper[4848]: E1204 13:49:04.538491 4848 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:49:04 crc kubenswrapper[4848]: E1204 13:49:04.538600 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs podName:5577aeec-1fac-4adc-a566-b0110dd2477b nodeName:}" failed. No retries permitted until 2025-12-04 13:49:12.53856608 +0000 UTC m=+56.481062648 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs") pod "network-metrics-daemon-v4k6x" (UID: "5577aeec-1fac-4adc-a566-b0110dd2477b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.604282 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.604329 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.604340 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.604358 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.604370 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.708028 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.708117 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.708130 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.708157 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.708173 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.811089 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.811160 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.811178 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.811206 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.811224 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.914771 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.914835 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.914854 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.914879 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:04 crc kubenswrapper[4848]: I1204 13:49:04.914898 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:04Z","lastTransitionTime":"2025-12-04T13:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.019213 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.019307 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.019334 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.019368 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.019393 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.122710 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.122781 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.122799 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.122829 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.122851 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.226762 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.226851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.226866 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.226889 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.226905 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.329743 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.329790 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.329826 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.329846 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.329865 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.433043 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.433106 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.433123 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.433150 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.433168 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.536471 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.536589 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.536610 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.536640 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.536663 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.638874 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.638942 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.638996 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.639029 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.639046 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.741930 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.741985 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.741993 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.742010 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.742019 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.845742 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.845826 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.845854 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.845889 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.845910 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.949649 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.949707 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.949723 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.949744 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:05 crc kubenswrapper[4848]: I1204 13:49:05.949756 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:05Z","lastTransitionTime":"2025-12-04T13:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.051993 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.052055 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.052076 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.052102 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.052122 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.154812 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.154878 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.154896 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.154921 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.154939 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.258742 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.258837 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.258889 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.258924 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.258999 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.362077 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.362153 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.362178 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.362210 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.362234 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.392498 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:06 crc kubenswrapper[4848]: E1204 13:49:06.392606 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.392749 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.392783 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:06 crc kubenswrapper[4848]: E1204 13:49:06.392809 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:06 crc kubenswrapper[4848]: E1204 13:49:06.392934 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.393020 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:06 crc kubenswrapper[4848]: E1204 13:49:06.393183 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.415159 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.435450 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.467636 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.467772 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.467797 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.467827 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.467915 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.482320 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:140\\\\nI1204 13:48:57.485444 6286 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:57.485459 6286 factory.go:656] Stopping watch factory\\\\nI1204 13:48:57.485470 6286 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:57.485478 6286 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:57.485484 6286 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:57.485529 6286 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:57.485843 6286 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486419 6286 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486852 6286 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486974 6286 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.487034 6286 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.506564 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.526569 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.560305 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.575992 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.576075 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.576094 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.576122 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.576147 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.582001 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.603395 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.623724 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.639399 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.658436 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.675322 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.680026 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.680112 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.680131 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.680157 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.680176 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.693461 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.706158 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.728249 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.743529 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.758920 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:06Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.783984 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.784044 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.784066 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.784097 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.784119 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.887502 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.887573 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.887591 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.887619 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.887637 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.990701 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.990782 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.990801 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.990834 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:06 crc kubenswrapper[4848]: I1204 13:49:06.990852 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:06Z","lastTransitionTime":"2025-12-04T13:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.093790 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.093850 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.093866 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.093890 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.093907 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.197489 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.197525 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.197537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.197556 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.197567 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.300491 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.300567 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.300585 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.300611 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.300629 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.403624 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.403679 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.403697 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.403723 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.403740 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.506532 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.506639 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.506656 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.506681 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.506698 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.572289 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.572518 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:49:39.572476917 +0000 UTC m=+83.514973475 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.572727 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.572842 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.573022 4848 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.573057 4848 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.573118 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:49:39.573097762 +0000 UTC m=+83.515594330 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.573171 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:49:39.573141453 +0000 UTC m=+83.515638021 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.610097 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.610159 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.610178 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.610203 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.610223 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.674395 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.674480 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.674579 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.674605 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.674620 4848 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.674618 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.674643 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.674661 4848 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.674684 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:49:39.674667503 +0000 UTC m=+83.617164031 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:49:07 crc kubenswrapper[4848]: E1204 13:49:07.674710 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:49:39.674693143 +0000 UTC m=+83.617189711 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.713237 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.713307 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.713326 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.713355 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.713375 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.817456 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.817527 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.817544 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.817569 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.817587 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.919821 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.919875 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.919892 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.919916 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:07 crc kubenswrapper[4848]: I1204 13:49:07.919933 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:07Z","lastTransitionTime":"2025-12-04T13:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.023537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.023593 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.023610 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.023636 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.023656 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.126831 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.126898 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.126916 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.126941 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.126989 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.232805 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.232915 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.232937 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.232996 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.233022 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.336001 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.336076 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.336094 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.336117 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.336135 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.392434 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.392510 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.392444 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:08 crc kubenswrapper[4848]: E1204 13:49:08.392632 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.392666 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:08 crc kubenswrapper[4848]: E1204 13:49:08.392803 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:08 crc kubenswrapper[4848]: E1204 13:49:08.393039 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:08 crc kubenswrapper[4848]: E1204 13:49:08.393167 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.439429 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.439506 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.439530 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.439560 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.439581 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.543074 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.543166 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.543192 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.543224 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.543247 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.646642 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.646834 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.646896 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.646927 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.647018 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.750530 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.750597 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.750615 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.750645 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.750673 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.853556 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.853640 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.853665 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.853697 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.853721 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.956703 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.956782 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.956822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.956850 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:08 crc kubenswrapper[4848]: I1204 13:49:08.956867 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:08Z","lastTransitionTime":"2025-12-04T13:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.059843 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.059938 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.060031 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.060059 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.060112 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.162977 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.163041 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.163059 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.163090 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.163112 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.258905 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.266536 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.267307 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.267343 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.267363 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.267373 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.272018 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.275495 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.295637 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.312832 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.329763 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.346903 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.364223 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.370293 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.370358 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.370375 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.370403 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.370423 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.378764 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.404375 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.425520 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.444482 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.464714 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.473563 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.473603 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.473615 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.473632 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.473644 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.482547 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.506211 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.525036 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.544467 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.576726 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.576794 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.576817 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.576849 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.576871 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.577241 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:140\\\\nI1204 13:48:57.485444 6286 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:57.485459 6286 factory.go:656] Stopping watch factory\\\\nI1204 13:48:57.485470 6286 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:57.485478 6286 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:57.485484 6286 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:57.485529 6286 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:57.485843 6286 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486419 6286 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486852 6286 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486974 6286 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.487034 6286 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.603072 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.680080 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.680128 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.680148 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.680174 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.680193 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.759042 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.759094 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.759110 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.759158 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.759175 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: E1204 13:49:09.780868 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.786443 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.786491 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.786509 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.786532 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.786550 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: E1204 13:49:09.808233 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.813427 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.813497 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.813518 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.813546 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.813564 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: E1204 13:49:09.834246 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.839998 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.840054 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.840072 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.840096 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.840112 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: E1204 13:49:09.862036 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.867630 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.867687 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.867707 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.867735 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.867752 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: E1204 13:49:09.888031 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:09Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:09 crc kubenswrapper[4848]: E1204 13:49:09.888255 4848 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.890356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.890422 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.890441 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.890467 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.890484 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.993718 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.993799 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.993820 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.993851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:09 crc kubenswrapper[4848]: I1204 13:49:09.993869 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:09Z","lastTransitionTime":"2025-12-04T13:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.096804 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.096919 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.096940 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.097010 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.097029 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.199890 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.199989 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.200015 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.200044 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.200068 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.302321 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.302381 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.302402 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.302427 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.302445 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.392920 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.393011 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:10 crc kubenswrapper[4848]: E1204 13:49:10.393067 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:10 crc kubenswrapper[4848]: E1204 13:49:10.393250 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.393314 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.393607 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.393972 4848 scope.go:117] "RemoveContainer" containerID="3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b" Dec 04 13:49:10 crc kubenswrapper[4848]: E1204 13:49:10.394021 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:10 crc kubenswrapper[4848]: E1204 13:49:10.393883 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.405574 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.405656 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.405673 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.405696 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.405717 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.509221 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.509587 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.509605 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.509629 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.509650 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.612669 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.612707 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.612727 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.612753 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.612771 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.716079 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.716122 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.716142 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.716165 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.716184 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.777472 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/1.log" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.780856 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.781507 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.806562 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.819573 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.819628 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.819656 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.819685 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.819704 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.836678 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.881867 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:140\\\\nI1204 13:48:57.485444 6286 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:57.485459 6286 factory.go:656] Stopping watch factory\\\\nI1204 13:48:57.485470 6286 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:57.485478 6286 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:57.485484 6286 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:57.485529 6286 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:57.485843 6286 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486419 6286 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486852 6286 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486974 6286 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.487034 6286 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.902671 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.920512 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.923218 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.923262 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.923274 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.923290 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.923304 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:10Z","lastTransitionTime":"2025-12-04T13:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.935354 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.960912 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.977142 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.988634 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:10 crc kubenswrapper[4848]: I1204 13:49:10.997856 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:10Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.006938 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.015341 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.025140 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.025371 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.025407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.025418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.025435 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.025446 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.035621 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.046985 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.059387 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.070067 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.079308 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.128067 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.128114 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.128124 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.128140 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.128152 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.231294 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.231363 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.231373 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.231389 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.231401 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.334548 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.334619 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.334637 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.334665 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.334686 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.437192 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.437268 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.437291 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.437324 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.437348 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.539539 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.539614 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.539623 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.539638 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.539647 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.643856 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.643929 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.643955 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.644016 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.644034 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.747806 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.747880 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.747898 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.747925 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.747945 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.788086 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/2.log" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.789061 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/1.log" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.793471 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2" exitCode=1 Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.793559 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.793661 4848 scope.go:117] "RemoveContainer" containerID="3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.794774 4848 scope.go:117] "RemoveContainer" containerID="cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2" Dec 04 13:49:11 crc kubenswrapper[4848]: E1204 13:49:11.795102 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.813636 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.834432 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.851799 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.851849 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.851861 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.851884 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.851900 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.852728 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.885808 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.901003 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.913192 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.926124 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.938765 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.953291 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.955340 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.955484 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.955510 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.955545 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.955569 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:11Z","lastTransitionTime":"2025-12-04T13:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.966174 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:11 crc kubenswrapper[4848]: I1204 13:49:11.985179 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:11Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.003738 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.017566 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.031894 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.059418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.059479 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.059497 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.059538 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.059563 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.059695 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b91a77dde925673c56b939d988f4ae24678edbd281a77e353d35aaa0cc9667b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"message\\\":\\\"ternalversions/factory.go:140\\\\nI1204 13:48:57.485444 6286 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 13:48:57.485459 6286 factory.go:656] Stopping watch factory\\\\nI1204 13:48:57.485470 6286 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 13:48:57.485478 6286 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 13:48:57.485484 6286 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 13:48:57.485529 6286 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 13:48:57.485843 6286 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486419 6286 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486852 6286 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.486974 6286 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 13:48:57.487034 6286 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.075846 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.094603 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.110643 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.162769 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.163145 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.163285 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.163439 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.163580 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.267080 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.267141 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.267157 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.267190 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.267204 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.371063 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.371133 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.371149 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.371175 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.371195 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.392541 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:12 crc kubenswrapper[4848]: E1204 13:49:12.392709 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.392824 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:12 crc kubenswrapper[4848]: E1204 13:49:12.392904 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.393027 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.393188 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:12 crc kubenswrapper[4848]: E1204 13:49:12.393273 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:12 crc kubenswrapper[4848]: E1204 13:49:12.393389 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.491568 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.491667 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.491689 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.491728 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.491745 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.595417 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.595474 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.595490 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.595514 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.595528 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.636921 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:12 crc kubenswrapper[4848]: E1204 13:49:12.637260 4848 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:49:12 crc kubenswrapper[4848]: E1204 13:49:12.637414 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs podName:5577aeec-1fac-4adc-a566-b0110dd2477b nodeName:}" failed. No retries permitted until 2025-12-04 13:49:28.637376877 +0000 UTC m=+72.579873445 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs") pod "network-metrics-daemon-v4k6x" (UID: "5577aeec-1fac-4adc-a566-b0110dd2477b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.699072 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.699142 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.699165 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.699199 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.699222 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.799852 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/2.log" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.801828 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.801877 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.801892 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.801911 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.801922 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.806033 4848 scope.go:117] "RemoveContainer" containerID="cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2" Dec 04 13:49:12 crc kubenswrapper[4848]: E1204 13:49:12.806347 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.822169 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.836620 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.861708 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.895032 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.904412 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.904459 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.904470 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.904488 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.904499 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:12Z","lastTransitionTime":"2025-12-04T13:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.914124 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.932294 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.946518 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.963652 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.975042 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:12 crc kubenswrapper[4848]: I1204 13:49:12.989858 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:12Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.006651 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.007832 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.007889 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.007934 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.008003 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.008022 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.019493 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.041182 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.057383 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.071602 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.084985 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.096616 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.109178 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:13Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.110594 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.110672 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.110693 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.110715 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.110731 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.213902 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.214231 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.214420 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.214564 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.214751 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.318355 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.318430 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.318453 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.318483 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.318504 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.422042 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.422116 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.422140 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.422168 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.422187 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.525471 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.525516 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.525565 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.525585 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.525596 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.628871 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.628916 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.628926 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.628978 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.628989 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.732691 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.732755 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.732771 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.732797 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.732815 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.836394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.836475 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.836495 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.836517 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.836535 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.939851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.939891 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.939905 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.939924 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:13 crc kubenswrapper[4848]: I1204 13:49:13.939937 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:13Z","lastTransitionTime":"2025-12-04T13:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.043131 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.043202 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.043220 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.043249 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.043269 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.146404 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.146448 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.146457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.146477 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.146487 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.249130 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.249180 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.249190 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.249207 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.249218 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.351498 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.351559 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.351579 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.351604 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.351620 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.392584 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.392664 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:14 crc kubenswrapper[4848]: E1204 13:49:14.392758 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:14 crc kubenswrapper[4848]: E1204 13:49:14.392907 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.393000 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.393026 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:14 crc kubenswrapper[4848]: E1204 13:49:14.393268 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:14 crc kubenswrapper[4848]: E1204 13:49:14.393390 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.453847 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.453895 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.453912 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.453934 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.453981 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.556685 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.556719 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.556729 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.556749 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.556762 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.659426 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.659491 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.659508 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.659533 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.659550 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.762587 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.762634 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.762649 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.762670 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.762684 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.865463 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.865518 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.865532 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.865551 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.865565 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.968710 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.968753 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.968764 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.968784 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:14 crc kubenswrapper[4848]: I1204 13:49:14.968798 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:14Z","lastTransitionTime":"2025-12-04T13:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.071689 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.071796 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.071817 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.071844 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.071862 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.175266 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.175333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.175377 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.175405 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.175423 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.277843 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.277887 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.277897 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.277913 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.277922 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.381301 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.381362 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.381385 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.381414 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.381439 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.483782 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.483829 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.483841 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.483857 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.483869 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.587194 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.587333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.587352 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.587383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.587403 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.689849 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.689913 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.689931 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.690022 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.690049 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.794135 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.794418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.794542 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.794835 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.794919 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.897523 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.897584 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.897596 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.897616 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:15 crc kubenswrapper[4848]: I1204 13:49:15.897630 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:15Z","lastTransitionTime":"2025-12-04T13:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.000039 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.000094 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.000111 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.000137 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.000156 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.103348 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.103412 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.103435 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.103470 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.103495 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.207042 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.207132 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.207158 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.207189 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.207207 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.309314 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.309370 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.309383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.309407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.309420 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.392447 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.392507 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.392484 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:16 crc kubenswrapper[4848]: E1204 13:49:16.392632 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:16 crc kubenswrapper[4848]: E1204 13:49:16.392709 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.392729 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:16 crc kubenswrapper[4848]: E1204 13:49:16.392878 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:16 crc kubenswrapper[4848]: E1204 13:49:16.393040 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.410504 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.412005 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.412066 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.412084 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.412113 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.412131 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.428768 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.447266 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.464588 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.479464 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.494181 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.514147 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.514181 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.514189 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.514203 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.514212 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.519326 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.532394 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.546555 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.556738 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.568601 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.579468 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.589435 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.601437 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.613897 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.616921 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.616993 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.617007 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.617023 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.617036 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.626780 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.640281 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.660783 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:16Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.720235 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.720280 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.720293 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.720310 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.720323 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.822734 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.822812 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.822830 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.822855 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.822874 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.926297 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.926409 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.926427 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.926451 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:16 crc kubenswrapper[4848]: I1204 13:49:16.926470 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:16Z","lastTransitionTime":"2025-12-04T13:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.030588 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.030686 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.030709 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.030746 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.030767 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.134201 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.134267 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.134291 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.134333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.134390 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.238085 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.238147 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.238165 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.238192 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.238210 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.341359 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.341412 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.341429 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.341457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.341474 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.444778 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.444841 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.444859 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.444882 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.444899 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.556528 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.556581 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.556597 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.556620 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.556633 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.658746 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.658802 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.658822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.658845 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.658865 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.762041 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.762089 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.762106 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.762128 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.762141 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.865824 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.865868 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.865880 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.865896 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.865907 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.969394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.969461 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.969481 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.969510 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:17 crc kubenswrapper[4848]: I1204 13:49:17.969528 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:17Z","lastTransitionTime":"2025-12-04T13:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.073866 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.074110 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.074136 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.074202 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.074225 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.177394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.177458 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.177474 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.177498 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.177515 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.280486 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.280543 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.280565 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.280589 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.280608 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.383817 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.383878 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.383895 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.383918 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.383938 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.392537 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.392562 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.392591 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.392565 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:18 crc kubenswrapper[4848]: E1204 13:49:18.392703 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:18 crc kubenswrapper[4848]: E1204 13:49:18.392826 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:18 crc kubenswrapper[4848]: E1204 13:49:18.392912 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:18 crc kubenswrapper[4848]: E1204 13:49:18.393021 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.486560 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.486598 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.486607 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.486622 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.486630 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.590579 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.592697 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.592751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.592780 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.592812 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.696053 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.696151 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.696168 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.696194 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.696213 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.799845 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.799995 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.800036 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.800068 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.800091 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.903549 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.903603 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.903616 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.903638 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:18 crc kubenswrapper[4848]: I1204 13:49:18.903652 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:18Z","lastTransitionTime":"2025-12-04T13:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.016443 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.016501 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.016518 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.016539 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.016556 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.120287 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.120352 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.120369 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.120394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.120414 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.223305 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.223366 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.223377 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.223396 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.223409 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.327395 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.327462 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.327480 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.327509 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.327526 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.430727 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.430795 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.430817 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.430851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.430870 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.533728 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.533795 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.533809 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.533840 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.533874 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.636493 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.636532 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.636557 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.636572 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.636583 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.738821 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.738865 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.738877 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.738894 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.738906 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.841649 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.841694 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.841710 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.841732 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.841746 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.945167 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.945237 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.945259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.945290 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:19 crc kubenswrapper[4848]: I1204 13:49:19.945312 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:19Z","lastTransitionTime":"2025-12-04T13:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.048604 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.048654 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.048672 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.048697 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.048716 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.101019 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.101061 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.101072 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.101089 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.101101 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.121301 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:20Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.125434 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.125531 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.125548 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.125566 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.125577 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.137275 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:20Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.140595 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.140636 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.140648 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.140665 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.140678 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.155015 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:20Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.158549 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.158588 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.158601 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.158617 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.158629 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.172801 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:20Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.176640 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.176682 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.176693 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.176709 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.176719 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.191175 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:20Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.191337 4848 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.193553 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.193591 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.193605 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.193624 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.193639 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.296692 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.296739 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.296756 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.296781 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.296801 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.392602 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.392742 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.392983 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.393060 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.393207 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.393276 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.393424 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:20 crc kubenswrapper[4848]: E1204 13:49:20.393496 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.400716 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.400753 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.400761 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.400775 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.400784 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.405480 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.507997 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.508037 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.508049 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.508069 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.508081 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.616129 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.616193 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.616212 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.616236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.616253 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.721858 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.721912 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.721924 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.721940 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.721969 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.824751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.824780 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.824789 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.824803 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.824812 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.927271 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.927333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.927348 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.927368 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:20 crc kubenswrapper[4848]: I1204 13:49:20.927385 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:20Z","lastTransitionTime":"2025-12-04T13:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.030434 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.030477 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.030489 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.030509 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.030522 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.134488 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.134544 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.134556 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.134576 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.134589 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.237813 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.237877 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.237903 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.237939 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.238000 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.341030 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.341097 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.341114 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.341139 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.341158 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.444693 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.444747 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.444764 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.444788 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.444808 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.547158 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.547211 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.547228 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.547251 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.547270 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.650764 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.650822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.650833 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.650856 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.650872 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.753680 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.753746 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.753765 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.753789 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.753806 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.858487 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.858572 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.858589 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.858616 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.858636 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.961816 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.961893 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.961909 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.961935 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:21 crc kubenswrapper[4848]: I1204 13:49:21.961982 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:21Z","lastTransitionTime":"2025-12-04T13:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.065589 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.065656 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.065668 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.065687 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.065700 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.168724 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.168806 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.168824 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.168854 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.168876 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.271837 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.271907 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.271924 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.271969 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.271986 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.375603 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.375667 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.375681 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.375705 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.375721 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.393050 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.393128 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.393183 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.393209 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:22 crc kubenswrapper[4848]: E1204 13:49:22.393345 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:22 crc kubenswrapper[4848]: E1204 13:49:22.393686 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:22 crc kubenswrapper[4848]: E1204 13:49:22.393860 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:22 crc kubenswrapper[4848]: E1204 13:49:22.394058 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.479510 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.479577 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.479587 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.479604 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.479616 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.583093 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.583147 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.583159 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.583180 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.583193 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.686945 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.687008 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.687018 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.687038 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.687050 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.790903 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.790989 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.790998 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.791013 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.791022 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.893575 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.893640 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.893655 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.893678 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.893694 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.997134 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.997203 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.997224 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.997262 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:22 crc kubenswrapper[4848]: I1204 13:49:22.997300 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:22Z","lastTransitionTime":"2025-12-04T13:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.100694 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.100767 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.100792 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.100832 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.100855 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.204195 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.204261 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.204280 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.204313 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.204331 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.307690 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.307755 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.307767 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.307786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.307800 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.411497 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.411550 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.411564 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.411585 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.411596 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.514030 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.514075 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.514084 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.514101 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.514112 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.617670 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.617732 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.617742 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.617760 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.617772 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.721383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.721473 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.721498 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.721533 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.721555 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.824653 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.824723 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.824735 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.824751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.824761 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.928562 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.928658 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.928675 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.928698 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:23 crc kubenswrapper[4848]: I1204 13:49:23.928714 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:23Z","lastTransitionTime":"2025-12-04T13:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.031690 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.031737 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.031748 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.031768 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.031780 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.134496 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.134532 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.134543 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.134558 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.134571 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.236823 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.236866 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.236910 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.236926 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.236935 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.340810 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.340865 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.340877 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.340893 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.340904 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.394074 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:24 crc kubenswrapper[4848]: E1204 13:49:24.394203 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.394291 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.394334 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.394364 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:24 crc kubenswrapper[4848]: E1204 13:49:24.394703 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:24 crc kubenswrapper[4848]: E1204 13:49:24.394847 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:24 crc kubenswrapper[4848]: E1204 13:49:24.394992 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.443580 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.443612 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.443622 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.443637 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.443649 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.545337 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.545366 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.545375 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.545386 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.545395 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.648654 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.648695 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.648706 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.648720 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.648731 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.752666 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.753017 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.753050 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.753077 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.753097 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.855115 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.855174 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.855199 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.855227 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.855250 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.958341 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.958371 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.958378 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.958391 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:24 crc kubenswrapper[4848]: I1204 13:49:24.958399 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:24Z","lastTransitionTime":"2025-12-04T13:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.061327 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.061371 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.061380 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.061394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.061404 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.165166 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.165217 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.165228 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.165249 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.165262 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.268715 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.268762 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.268771 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.268786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.268799 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.371523 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.371559 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.371568 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.371583 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.371592 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.474162 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.474236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.474259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.474286 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.474307 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.577155 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.577197 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.577208 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.577227 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.577238 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.679819 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.679890 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.679916 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.679983 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.680010 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.783579 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.783637 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.783656 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.783679 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.783696 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.887929 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.888036 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.888064 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.888096 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.888119 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.992097 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.992155 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.992169 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.992194 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:25 crc kubenswrapper[4848]: I1204 13:49:25.992210 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:25Z","lastTransitionTime":"2025-12-04T13:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.095750 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.095803 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.095819 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.095842 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.095855 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.207629 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.207683 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.207695 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.207717 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.207731 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.311123 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.311176 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.311189 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.311207 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.311219 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.392636 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.392839 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.392756 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:26 crc kubenswrapper[4848]: E1204 13:49:26.393337 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:26 crc kubenswrapper[4848]: E1204 13:49:26.393479 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.393513 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:26 crc kubenswrapper[4848]: E1204 13:49:26.393187 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:26 crc kubenswrapper[4848]: E1204 13:49:26.393647 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.407699 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.413748 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.413777 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.413786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.413798 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.413807 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.418374 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.438905 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.453844 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.465757 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.488119 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.499068 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd70a0d0-5261-49f7-bcac-ddba55df45aa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.516498 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.516701 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.516740 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.516753 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.516774 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.516789 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.539626 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.553410 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.572802 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.589306 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.604999 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.620169 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.620216 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.620228 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.620246 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.620261 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.633089 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.653376 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.669175 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.686208 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.699406 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.714839 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:26Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.721766 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.721802 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.721812 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.721827 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.721839 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.824491 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.824537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.824547 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.824562 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.824571 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.927647 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.927722 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.927745 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.927775 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:26 crc kubenswrapper[4848]: I1204 13:49:26.927797 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:26Z","lastTransitionTime":"2025-12-04T13:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.029689 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.029763 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.029814 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.029850 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.029871 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.135806 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.135843 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.135852 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.135866 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.135878 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.238690 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.238745 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.238757 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.238778 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.238789 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.341893 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.341997 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.342066 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.342089 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.342107 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.394130 4848 scope.go:117] "RemoveContainer" containerID="cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2" Dec 04 13:49:27 crc kubenswrapper[4848]: E1204 13:49:27.394433 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.445251 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.445366 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.445385 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.445407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.445424 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.548799 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.548847 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.548864 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.548888 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.548903 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.651496 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.651525 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.651533 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.651547 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.651556 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.753899 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.754012 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.754088 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.754123 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.754146 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.858876 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.858923 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.858941 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.859005 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.859025 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.962715 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.962768 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.962784 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.962809 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:27 crc kubenswrapper[4848]: I1204 13:49:27.962827 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:27Z","lastTransitionTime":"2025-12-04T13:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.066204 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.066248 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.066260 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.066282 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.066294 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.169483 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.169547 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.169563 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.169592 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.169611 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.272874 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.272975 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.273002 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.273033 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.273053 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.376165 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.376561 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.376731 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.376900 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.377127 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.392603 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.392619 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.392880 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.392673 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:28 crc kubenswrapper[4848]: E1204 13:49:28.393246 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:28 crc kubenswrapper[4848]: E1204 13:49:28.393414 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:28 crc kubenswrapper[4848]: E1204 13:49:28.393521 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:28 crc kubenswrapper[4848]: E1204 13:49:28.393571 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.480425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.480510 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.480536 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.480566 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.480587 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.583990 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.584056 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.584079 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.584102 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.584119 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.686324 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.686394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.686433 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.686465 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.686486 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.736082 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:28 crc kubenswrapper[4848]: E1204 13:49:28.736530 4848 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:49:28 crc kubenswrapper[4848]: E1204 13:49:28.736809 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs podName:5577aeec-1fac-4adc-a566-b0110dd2477b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:00.736777491 +0000 UTC m=+104.679274059 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs") pod "network-metrics-daemon-v4k6x" (UID: "5577aeec-1fac-4adc-a566-b0110dd2477b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.789287 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.789320 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.789328 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.789341 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.789350 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.891642 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.891686 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.891694 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.891709 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.891723 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.993856 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.993925 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.993944 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.994002 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:28 crc kubenswrapper[4848]: I1204 13:49:28.994022 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:28Z","lastTransitionTime":"2025-12-04T13:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.097407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.097480 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.097503 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.097531 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.097552 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.200923 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.201023 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.201039 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.201062 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.201083 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.304549 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.304624 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.304645 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.304671 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.304689 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.407519 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.407582 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.407603 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.407628 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.407649 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.510595 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.510659 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.510682 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.510710 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.510730 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.612570 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.612602 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.612611 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.612625 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.612633 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.719528 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.719568 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.719580 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.719595 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.719606 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.821727 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.821800 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.821822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.821851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.821870 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.924924 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.925008 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.925027 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.925052 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:29 crc kubenswrapper[4848]: I1204 13:49:29.925071 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:29Z","lastTransitionTime":"2025-12-04T13:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.028882 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.028993 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.029016 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.029041 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.029061 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.131865 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.131933 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.131980 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.132004 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.132020 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.234829 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.234930 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.235422 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.235708 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.236027 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.338599 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.338641 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.338657 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.338679 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.338696 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.393135 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.397448 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.397653 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.397647 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.397827 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.398038 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.398081 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.398134 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.441098 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.441141 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.441153 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.441167 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.441181 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.445439 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.445493 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.445510 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.445529 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.445561 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.465749 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.469281 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.469322 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.469334 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.469352 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.469365 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.483069 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.486376 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.486418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.486431 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.486446 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.486459 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.500165 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.503933 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.504015 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.504027 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.504047 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.504061 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.518656 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.522823 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.522866 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.522884 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.522907 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.522924 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.536672 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: E1204 13:49:30.536879 4848 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.543413 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.543467 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.543482 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.543505 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.543521 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.646112 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.646160 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.646176 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.646198 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.646213 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.748612 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.748652 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.748661 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.748679 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.748693 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.852033 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.852077 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.852093 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.852115 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.852132 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.871107 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/0.log" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.871169 4848 generic.go:334] "Generic (PLEG): container finished" podID="2b6bff84-ab72-4936-8c9f-0508ded8fdd8" containerID="a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99" exitCode=1 Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.871203 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spvsd" event={"ID":"2b6bff84-ab72-4936-8c9f-0508ded8fdd8","Type":"ContainerDied","Data":"a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.871655 4848 scope.go:117] "RemoveContainer" containerID="a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.889763 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.900883 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.914886 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.931753 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"2025-12-04T13:48:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50\\\\n2025-12-04T13:48:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50 to /host/opt/cni/bin/\\\\n2025-12-04T13:48:45Z [verbose] multus-daemon started\\\\n2025-12-04T13:48:45Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:49:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.945933 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.956832 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.956904 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.956915 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.956992 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.957027 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:30Z","lastTransitionTime":"2025-12-04T13:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.964584 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.978592 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd70a0d0-5261-49f7-bcac-ddba55df45aa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:30 crc kubenswrapper[4848]: I1204 13:49:30.997804 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:30Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.024650 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.036563 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.054834 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.060033 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.060080 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.060091 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.060109 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.060122 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.066403 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.094396 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.105090 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.119852 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.130186 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.138445 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.149892 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.162030 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.162124 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.162143 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.162169 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.162185 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.164278 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.265418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.265481 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.265501 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.265523 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.265540 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.368371 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.368425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.368437 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.368456 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.368468 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.470892 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.470932 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.470942 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.470975 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.470987 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.573860 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.573937 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.574026 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.574059 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.574086 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.677396 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.677461 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.677478 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.677502 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.677521 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.780326 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.780383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.780406 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.780434 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.780455 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.877030 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/0.log" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.877096 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spvsd" event={"ID":"2b6bff84-ab72-4936-8c9f-0508ded8fdd8","Type":"ContainerStarted","Data":"5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.882712 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.882767 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.882782 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.882804 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.882820 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.895514 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"2025-12-04T13:48:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50\\\\n2025-12-04T13:48:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50 to /host/opt/cni/bin/\\\\n2025-12-04T13:48:45Z [verbose] multus-daemon started\\\\n2025-12-04T13:48:45Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:49:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.912844 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.930975 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.945785 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.967501 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.985456 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.985516 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.985528 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.985546 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.985557 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:31Z","lastTransitionTime":"2025-12-04T13:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:31 crc kubenswrapper[4848]: I1204 13:49:31.991584 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:31Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.011256 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.024183 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd70a0d0-5261-49f7-bcac-ddba55df45aa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.042617 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.053980 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.066400 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.078462 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.088233 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.088256 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.088266 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.088283 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.088295 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.091982 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.102657 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.113772 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.124863 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.150255 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.163235 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.175220 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:32Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.190967 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.191001 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.191009 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.191024 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.191034 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.293669 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.293722 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.293738 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.293759 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.293784 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.393555 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.393635 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.393567 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:32 crc kubenswrapper[4848]: E1204 13:49:32.393807 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:32 crc kubenswrapper[4848]: E1204 13:49:32.393980 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:32 crc kubenswrapper[4848]: E1204 13:49:32.394134 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.394171 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:32 crc kubenswrapper[4848]: E1204 13:49:32.394246 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.395844 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.395906 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.395927 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.395981 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.396000 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.499273 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.499376 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.499392 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.499416 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.499434 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.602150 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.602204 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.602242 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.602260 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.602271 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.705628 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.705678 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.705689 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.705706 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.705717 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.808903 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.808990 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.809011 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.809033 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.809052 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.912267 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.912325 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.912341 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.912364 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:32 crc kubenswrapper[4848]: I1204 13:49:32.912382 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:32Z","lastTransitionTime":"2025-12-04T13:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.015048 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.015161 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.015188 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.015217 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.015240 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.118648 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.118719 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.118743 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.118770 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.118791 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.222018 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.222090 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.222124 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.222151 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.222171 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.325582 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.325663 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.325686 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.325711 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.325728 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.429188 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.429240 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.429259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.429280 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.429296 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.532381 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.532481 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.532499 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.532633 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.532737 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.635841 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.635895 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.635912 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.635935 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.635987 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.738758 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.738810 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.738828 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.738849 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.738866 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.842305 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.842405 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.842459 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.842483 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.842503 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.945576 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.945653 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.945668 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.945690 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:33 crc kubenswrapper[4848]: I1204 13:49:33.946036 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:33Z","lastTransitionTime":"2025-12-04T13:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.049569 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.049647 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.049665 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.049729 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.049759 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.153328 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.153383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.153394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.153418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.153430 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.256350 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.256397 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.256407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.256428 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.256463 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.363834 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.363900 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.363910 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.363929 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.363941 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.392457 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.392500 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.392457 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:34 crc kubenswrapper[4848]: E1204 13:49:34.392645 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.392669 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:34 crc kubenswrapper[4848]: E1204 13:49:34.392746 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:34 crc kubenswrapper[4848]: E1204 13:49:34.392816 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:34 crc kubenswrapper[4848]: E1204 13:49:34.392872 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.467411 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.467489 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.467508 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.467537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.467562 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.572700 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.572804 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.572822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.572847 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.572867 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.675863 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.675926 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.675941 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.675988 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.676004 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.778813 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.778879 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.778890 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.778908 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.778919 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.882712 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.882786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.882797 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.882822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.882835 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.986552 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.986722 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.986755 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.986782 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:34 crc kubenswrapper[4848]: I1204 13:49:34.986803 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:34Z","lastTransitionTime":"2025-12-04T13:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.091761 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.091854 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.091874 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.091898 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.091915 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.195822 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.195897 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.195922 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.196062 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.196160 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.299777 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.299914 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.300043 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.300075 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.300096 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.402774 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.402803 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.402821 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.402835 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.402846 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.506319 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.506374 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.506394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.506418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.506437 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.609693 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.610056 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.610290 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.610535 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.610928 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.715726 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.716116 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.716275 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.716421 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.716556 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.820284 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.820651 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.820889 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.821119 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.821312 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.923582 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.923622 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.923635 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.923655 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:35 crc kubenswrapper[4848]: I1204 13:49:35.923669 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:35Z","lastTransitionTime":"2025-12-04T13:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.028522 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.028996 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.029297 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.029605 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.029886 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.138074 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.138131 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.138150 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.138175 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.138239 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.241376 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.241423 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.241439 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.241461 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.241479 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.344304 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.344372 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.344390 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.344419 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.344438 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.393223 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.393368 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.393450 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.393638 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:36 crc kubenswrapper[4848]: E1204 13:49:36.395375 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:36 crc kubenswrapper[4848]: E1204 13:49:36.395452 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:36 crc kubenswrapper[4848]: E1204 13:49:36.395515 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:36 crc kubenswrapper[4848]: E1204 13:49:36.395563 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.411101 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.427306 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.444728 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.449146 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.449316 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.449538 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.449665 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.449767 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.463152 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.502928 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.526179 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.543409 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.553884 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.554133 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.554400 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.554609 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.554798 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.562640 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"2025-12-04T13:48:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50\\\\n2025-12-04T13:48:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50 to /host/opt/cni/bin/\\\\n2025-12-04T13:48:45Z [verbose] multus-daemon started\\\\n2025-12-04T13:48:45Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:49:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.580822 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.596332 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.616015 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.638074 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.663164 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.663277 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.663290 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.663307 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.663320 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.667257 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.683010 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.695992 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd70a0d0-5261-49f7-bcac-ddba55df45aa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.710610 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.741635 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.758351 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.768068 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.768113 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.768130 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.768153 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.768170 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.772418 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:36Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.870347 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.870391 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.870404 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.870422 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.870433 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.972877 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.972992 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.973015 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.973044 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:36 crc kubenswrapper[4848]: I1204 13:49:36.973067 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:36Z","lastTransitionTime":"2025-12-04T13:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.076510 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.076585 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.076604 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.076627 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.076644 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.180322 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.180385 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.180402 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.180426 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.180443 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.283337 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.283404 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.283421 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.283446 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.283463 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.386430 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.386492 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.386513 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.386542 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.386563 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.490497 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.490559 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.490576 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.490599 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.490619 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.593685 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.593743 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.593763 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.593787 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.593806 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.697481 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.698009 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.698053 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.698135 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.698478 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.802591 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.802650 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.802674 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.802702 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.802720 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.906220 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.906408 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.906434 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.906461 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:37 crc kubenswrapper[4848]: I1204 13:49:37.906488 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:37Z","lastTransitionTime":"2025-12-04T13:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.011100 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.011145 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.011156 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.011170 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.011180 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.114200 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.114240 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.114249 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.114263 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.114272 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.217436 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.217507 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.217527 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.217665 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.217741 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.321131 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.321214 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.321235 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.321267 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.321292 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.393468 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.393998 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.394070 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.394211 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:38 crc kubenswrapper[4848]: E1204 13:49:38.394221 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:38 crc kubenswrapper[4848]: E1204 13:49:38.394351 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:38 crc kubenswrapper[4848]: E1204 13:49:38.394476 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.394498 4848 scope.go:117] "RemoveContainer" containerID="cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2" Dec 04 13:49:38 crc kubenswrapper[4848]: E1204 13:49:38.394585 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.426263 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.426535 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.426736 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.426902 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.427113 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.529511 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.529575 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.529592 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.529617 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.529633 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.634244 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.634301 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.634318 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.634341 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.634359 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.737363 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.737425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.737442 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.737465 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.737483 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.840027 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.840097 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.840113 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.840138 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.840157 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.942812 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.943019 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.943037 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.943078 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:38 crc kubenswrapper[4848]: I1204 13:49:38.943095 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:38Z","lastTransitionTime":"2025-12-04T13:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.045547 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.045615 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.045632 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.045657 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.045674 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.149044 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.149097 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.149113 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.149136 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.149154 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.255264 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.255326 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.255383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.255403 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.255455 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.359785 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.359864 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.359876 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.359898 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.359910 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.463041 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.463088 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.463130 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.463152 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.463162 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.566025 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.566085 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.566096 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.566115 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.566125 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.661323 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.661482 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:43.661446068 +0000 UTC m=+147.603942606 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.661588 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.661679 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.661767 4848 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.661821 4848 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.661846 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:50:43.661826577 +0000 UTC m=+147.604323105 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.661865 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 13:50:43.661855858 +0000 UTC m=+147.604352406 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.668734 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.668773 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.668784 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.668804 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.668817 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.763021 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.763097 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.763242 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.763285 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.763297 4848 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.763353 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 13:50:43.763337951 +0000 UTC m=+147.705834479 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.763253 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.763396 4848 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.763411 4848 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:49:39 crc kubenswrapper[4848]: E1204 13:49:39.763463 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 13:50:43.763446474 +0000 UTC m=+147.705943112 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.771149 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.771186 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.771195 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.771210 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.771220 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.874727 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.874765 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.874776 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.874789 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.874799 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.917624 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/2.log" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.921528 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.922233 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.943894 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.960710 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd70a0d0-5261-49f7-bcac-ddba55df45aa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.977026 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.977071 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.977083 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.977100 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.977111 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:39Z","lastTransitionTime":"2025-12-04T13:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:39 crc kubenswrapper[4848]: I1204 13:49:39.980001 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:39Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.009652 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.026750 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.046744 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.064430 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.075739 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.079774 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.079814 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.079824 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.079841 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.079938 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.099209 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.116491 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.132491 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.147437 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.161158 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.174460 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.183088 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.183132 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.183149 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.183173 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.183188 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.188979 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.208897 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.233747 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.249527 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"2025-12-04T13:48:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50\\\\n2025-12-04T13:48:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50 to /host/opt/cni/bin/\\\\n2025-12-04T13:48:45Z [verbose] multus-daemon started\\\\n2025-12-04T13:48:45Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:49:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.264377 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.286640 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.286761 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.286840 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.286906 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.287013 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.390466 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.390532 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.390551 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.390577 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.390598 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.393095 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.393156 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.393156 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.393270 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.393268 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.393482 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.393629 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.393741 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.494579 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.494645 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.494662 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.494686 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.494705 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.596729 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.596808 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.596825 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.596845 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.596858 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.699706 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.699767 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.699792 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.699823 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.699846 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.803043 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.803112 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.803135 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.803166 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.803191 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.879934 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.880015 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.880032 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.880056 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.880076 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.902850 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.908375 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.908430 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.908446 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.908471 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.908489 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.928640 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/3.log" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.929817 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/2.log" Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.929663 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.935085 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.935161 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.935187 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.935218 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.935240 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.936631 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02" exitCode=1 Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.936692 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.936739 4848 scope.go:117] "RemoveContainer" containerID="cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.937919 4848 scope.go:117] "RemoveContainer" containerID="83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02" Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.938259 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.959100 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.964331 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.964388 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.964407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.964437 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.964460 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.965341 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"2025-12-04T13:48:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50\\\\n2025-12-04T13:48:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50 to /host/opt/cni/bin/\\\\n2025-12-04T13:48:45Z [verbose] multus-daemon started\\\\n2025-12-04T13:48:45Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:49:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: E1204 13:49:40.981152 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.982694 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.986116 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.986157 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.986169 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.986188 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.986205 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:40Z","lastTransitionTime":"2025-12-04T13:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:40 crc kubenswrapper[4848]: I1204 13:49:40.998322 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:40Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: E1204 13:49:41.002995 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: E1204 13:49:41.003322 4848 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.005118 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.005355 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.005511 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.005685 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.005870 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.011636 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.032037 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.058432 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd5b984ea2cca254cef4e4bd9c57a3e910ca1006667c80d08317b5f3363c27b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:11Z\\\",\\\"message\\\":\\\":,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.233],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 13:49:11.455215 6477 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456232 6477 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8rt9n\\\\nI1204 13:49:11.456249 6477 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8rt9n in node crc\\\\nI1204 13:49:11.456248 6477 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler-operator/metrics for network=default are: map[]\\\\nF1204 13:49:11.456255 6477 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network co\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"s:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 13:49:40.038389 6870 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 13:49:40.037400 6870 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:49:40.038549 6870 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:49:40.038597 6870 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1204 13:49:40.038668 6870 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.071797 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.082462 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd70a0d0-5261-49f7-bcac-ddba55df45aa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.096251 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.106883 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.108894 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.108933 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.108969 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.108989 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.109005 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.119662 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.129814 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.139784 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.149285 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.158850 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.167477 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.189732 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.205536 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.211254 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.211297 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.211309 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.211327 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.211342 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.218884 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.314350 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.314445 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.314468 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.314490 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.314508 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.416600 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.416639 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.416649 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.416662 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.416672 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.519490 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.519534 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.519545 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.519564 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.519576 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.622077 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.622142 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.622153 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.622169 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.622180 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.725120 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.725205 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.725229 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.725260 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.725279 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.829128 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.829226 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.829250 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.829278 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.829294 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.932243 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.932310 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.932329 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.932355 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.932375 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:41Z","lastTransitionTime":"2025-12-04T13:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.943693 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/3.log" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.950153 4848 scope.go:117] "RemoveContainer" containerID="83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02" Dec 04 13:49:41 crc kubenswrapper[4848]: E1204 13:49:41.950495 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:49:41 crc kubenswrapper[4848]: I1204 13:49:41.982395 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"s:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 13:49:40.038389 6870 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 13:49:40.037400 6870 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:49:40.038549 6870 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:49:40.038597 6870 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1204 13:49:40.038668 6870 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.002726 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:41Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.017474 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd70a0d0-5261-49f7-bcac-ddba55df45aa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.034578 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.034625 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.034638 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.034656 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.034668 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.038541 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.055026 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.078027 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.093929 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.111238 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.125299 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.138226 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.138319 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.138347 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.138381 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.138404 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.140117 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.157061 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.189167 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.206655 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.223345 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.241717 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.241757 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.241771 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.241790 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.241806 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.244459 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"2025-12-04T13:48:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50\\\\n2025-12-04T13:48:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50 to /host/opt/cni/bin/\\\\n2025-12-04T13:48:45Z [verbose] multus-daemon started\\\\n2025-12-04T13:48:45Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:49:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.258518 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.273374 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.290508 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.306332 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:42Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.345105 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.345166 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.345182 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.345206 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.345223 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.393227 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.393317 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.393436 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:42 crc kubenswrapper[4848]: E1204 13:49:42.393430 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:42 crc kubenswrapper[4848]: E1204 13:49:42.393564 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:42 crc kubenswrapper[4848]: E1204 13:49:42.393649 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.393822 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:42 crc kubenswrapper[4848]: E1204 13:49:42.394087 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.448078 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.448117 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.448127 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.448139 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.448148 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.550671 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.550749 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.550769 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.550795 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.550815 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.654181 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.654249 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.654259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.654273 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.654285 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.757169 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.757236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.757258 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.757285 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.757308 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.860154 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.860201 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.860210 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.860227 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.860238 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.964160 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.964220 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.964238 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.964262 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:42 crc kubenswrapper[4848]: I1204 13:49:42.964280 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:42Z","lastTransitionTime":"2025-12-04T13:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.066883 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.066924 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.066932 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.066962 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.066971 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.169763 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.169836 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.169849 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.169866 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.169877 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.273297 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.273346 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.273361 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.273381 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.273395 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.377370 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.377447 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.377465 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.377489 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.377509 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.481597 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.482734 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.483014 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.483259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.483479 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.586349 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.586418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.586436 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.586462 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.586481 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.690607 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.690668 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.690686 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.690710 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.690727 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.793176 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.793673 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.793874 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.794091 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.794318 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.897537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.898034 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.898240 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.898394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:43 crc kubenswrapper[4848]: I1204 13:49:43.898538 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:43Z","lastTransitionTime":"2025-12-04T13:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.000549 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.000602 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.000618 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.000639 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.000656 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.104075 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.104794 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.105023 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.105213 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.105456 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.209117 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.209163 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.209179 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.209196 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.209208 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.312123 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.312179 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.312197 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.312221 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.312238 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.393046 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.393113 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:44 crc kubenswrapper[4848]: E1204 13:49:44.393189 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.393046 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:44 crc kubenswrapper[4848]: E1204 13:49:44.393404 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:44 crc kubenswrapper[4848]: E1204 13:49:44.393520 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.393861 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:44 crc kubenswrapper[4848]: E1204 13:49:44.394045 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.414495 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.414577 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.414599 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.414624 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.414642 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.518413 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.518480 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.518498 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.518527 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.518546 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.621503 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.621578 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.621596 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.621619 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.621636 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.724909 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.724984 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.725001 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.725025 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.725043 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.828379 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.828454 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.828474 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.828504 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.828529 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.934559 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.934665 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.934700 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.934728 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:44 crc kubenswrapper[4848]: I1204 13:49:44.934747 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:44Z","lastTransitionTime":"2025-12-04T13:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.038861 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.039196 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.039281 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.039353 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.039441 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.143547 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.143612 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.143624 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.143647 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.143661 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.246823 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.246878 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.246896 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.246919 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.246935 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.349979 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.350022 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.350053 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.350067 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.350077 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.453336 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.453422 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.453449 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.453480 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.453502 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.557374 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.557417 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.557425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.557441 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.557450 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.660387 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.660437 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.660448 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.660466 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.660478 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.763341 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.763391 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.763407 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.763432 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.763449 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.866934 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.867006 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.867020 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.867037 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.867049 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.995300 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.995345 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.995357 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.995376 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:45 crc kubenswrapper[4848]: I1204 13:49:45.995389 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:45Z","lastTransitionTime":"2025-12-04T13:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.098104 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.098180 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.098198 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.098223 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.098240 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.201368 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.201421 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.201441 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.201464 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.201485 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.304717 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.304793 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.304815 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.304846 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.304870 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.393335 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:46 crc kubenswrapper[4848]: E1204 13:49:46.393521 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.393552 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.393598 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:46 crc kubenswrapper[4848]: E1204 13:49:46.393661 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:46 crc kubenswrapper[4848]: E1204 13:49:46.393739 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.393846 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:46 crc kubenswrapper[4848]: E1204 13:49:46.394077 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.408370 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.408405 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.408416 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.408430 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.408445 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.411647 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e95b5c9-2490-4d86-aff3-f255fa54ad67\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b86f2f45944ede700fa04ca265da60b621e21d74ea72adf43e50b36975183f08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d99ee331b50f04a196faafc3247c87ac46772d726cb67b989750068974c028\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589034199f99af48ccac8586df306ce79802d10407f7c3983f0808e37098ad0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.423525 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bedd115c-06c2-484b-88ab-4748c86b4531\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6235a4d8acc9f3a545a85d9a0df71dde65ca3eb164e07bc2a0485b2910e8c93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qm5tr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jspn5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.452400 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-56gg2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8a59239-bebd-48bc-85d5-1b2062f90d8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f373362422523443d587a61c0625d6c3bfcfe050ce2675fd7a50e038e1001e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8019151925df4674c32ef4881abd56d12bd553fecdfda29874163826a76eae4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3447ae27a1061a455d56b42881614280e0700fe5e7a00bc997f83f196c361017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38ee473ae6a12f92085298bfbc85cb144cf3cf62357d48723ac867d42dd5912e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce3d6e129ceb688648d551044f2918e25e862edc68a52d39d6cf0fb8303b4bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://480dfce268e4fe95c86133e5bd2dd21d88649c82e734458d3b08bb599b3aa8e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e0b0408a7656dd57aab34646070283fa44c8f9e49619aa56b0a1b64ea464f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nz2jc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-56gg2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.491132 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spvsd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6bff84-ab72-4936-8c9f-0508ded8fdd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:30Z\\\",\\\"message\\\":\\\"2025-12-04T13:48:45+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50\\\\n2025-12-04T13:48:45+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d27e3e39-7ed1-4228-9804-6d35a5697d50 to /host/opt/cni/bin/\\\\n2025-12-04T13:48:45Z [verbose] multus-daemon started\\\\n2025-12-04T13:48:45Z [verbose] Readiness Indicator file check\\\\n2025-12-04T13:49:30Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kth2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spvsd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:46Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.510421 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.510453 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.510463 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.510480 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.510490 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.612918 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.612982 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.612992 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.613007 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.613015 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.715093 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.715131 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.715140 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.715153 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.715162 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.817314 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.817374 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.817392 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.817419 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.817440 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.921158 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.921232 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.921249 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.921295 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:46 crc kubenswrapper[4848]: I1204 13:49:46.921313 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:46Z","lastTransitionTime":"2025-12-04T13:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.024772 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.024868 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.024894 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.024927 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.024991 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.128207 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.128285 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.128312 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.128344 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.128363 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.232003 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.232075 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.232092 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.232520 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.232586 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.336528 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.336580 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.336597 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.336619 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.336635 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.382179 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8rt9n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"599496f2-4367-4daf-9146-040bec1cd03d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f9628939c9a04ed1224f33dfb55e861309444b75909e028ec720bcc6e9ade8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2m7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:46Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8rt9n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.406050 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.423104 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd70a0d0-5261-49f7-bcac-ddba55df45aa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f7daccc2509f75819ae4940081616fac9d7c2c1411807514727a3541fbaa4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6e73f39c8fa3a6dc41f48a255f88e596e2dd0a4c2ffbc45a4159ac03562a8cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.440250 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.440313 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.440329 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.440353 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.440374 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.443889 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.468902 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88de7a37-3579-4630-9aa7-64654e68a472\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T13:49:40Z\\\",\\\"message\\\":\\\"s:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 13:49:40.038389 6870 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 13:49:40.037400 6870 ovnkube.go:599] Stopped ovnkube\\\\nI1204 13:49:40.038549 6870 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 13:49:40.038597 6870 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1204 13:49:40.038668 6870 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T13:49:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h7z6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-vwdkp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.489701 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.511235 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.532987 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.542356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.542425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.542446 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.542471 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.542489 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.572082 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f233e7b7-1c64-45c1-bf9c-34acae8651f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd7721730aa0b71309fabe8f95b44c4d85253366cb486fa45584ed79130aa638\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95c6a8b0c94be081fbf6a557af86fed5bdb7e2d70d9373523dc075fee2f77b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a02135a94c4254432e0bf0a35e133ed802eacdc3b2d1cb5f7fd43db58a09ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3135ee7a12ccb83966af5cb033ba725c07588606c055cc4861436338a93f3795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b57beb8499a169e7423e744a13e14404217051adb1e0b4b0aee6310891e9e866\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1824d6c871ed5d57e77c05191e2a355163f00e5e0dd3eb6cfe0c933fbe90ae73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29ceb11c744bbfb9b1cae1c3a07f6d60fe6a02a0b64d0aca18d5dfa3bfcdc37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d817ea09b10725f155fa90fcbfc657f1e6c7f6b1283a672aa2e737e0b244ad7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.589920 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.605349 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5abedfbb8c9f55bd35ce6daa3aa49bdef9254d000285776c9eebcf2d372fbdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97f4d387ce535e87c6ab743430b0995215891cfcfefb6c964df2df8e2d5a62bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.622783 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae40c568e2ce4d240529461dfe54c37873b0b6db62410485b2f92aa4368f8b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.637892 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lpwtz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b23ff2f-6e32-4b5c-9867-9988fa46d456\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b92324530a2d7bbd225a3608fc69e696ea47c5162f1464e89ca92ed3c2e0fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qd5g4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lpwtz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.644789 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.644836 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.644851 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.644876 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.644894 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.654930 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"937d7bdf-0181-4852-8d9e-1eb0a0540577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b87467aa5a599aeeb5faa7e17221ceae61ba96ecd92ab8d26abfebc4ee8fefe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d31395fb779d0e25d57e61c519f5f8f2400e18ab5d71342b3d5737de520500f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kg7cd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fd6s7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.668138 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:47Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.747445 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.747539 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.747562 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.747593 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.747616 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.850543 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.850639 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.850662 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.850689 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.850706 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.953714 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.953773 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.953790 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.953814 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:47 crc kubenswrapper[4848]: I1204 13:49:47.953829 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:47Z","lastTransitionTime":"2025-12-04T13:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.056226 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.056328 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.056356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.056384 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.056404 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.158922 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.159009 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.159027 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.159048 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.159065 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.263319 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.263385 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.263405 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.263432 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.263450 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.367165 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.367242 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.367265 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.367297 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.367321 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.392530 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.392614 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.392571 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:48 crc kubenswrapper[4848]: E1204 13:49:48.392785 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.392839 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:48 crc kubenswrapper[4848]: E1204 13:49:48.392929 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:48 crc kubenswrapper[4848]: E1204 13:49:48.393059 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:48 crc kubenswrapper[4848]: E1204 13:49:48.393749 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.470361 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.470418 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.470438 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.470461 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.470478 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.573893 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.573983 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.574003 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.574024 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.574041 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.677637 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.677699 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.677722 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.677751 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.677774 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.781205 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.781265 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.781330 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.781358 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.781376 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.884732 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.884798 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.884817 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.884842 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.884858 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.988094 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.988170 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.988193 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.988222 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:48 crc kubenswrapper[4848]: I1204 13:49:48.988244 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:48Z","lastTransitionTime":"2025-12-04T13:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.092022 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.092093 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.092115 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.092144 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.092165 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.195003 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.195053 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.195069 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.195093 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.195114 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.298123 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.298234 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.298253 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.298277 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.298294 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.401251 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.401325 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.401348 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.401373 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.401395 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.504365 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.504436 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.504459 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.504488 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.504509 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.607190 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.607251 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.607267 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.607287 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.607303 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.711018 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.711059 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.711075 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.711104 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.711119 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.813931 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.814001 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.814015 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.814031 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.814043 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.916809 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.916867 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.916878 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.916895 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:49 crc kubenswrapper[4848]: I1204 13:49:49.916908 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:49Z","lastTransitionTime":"2025-12-04T13:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.020137 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.020185 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.020194 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.020212 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.020225 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.123179 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.123236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.123253 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.123275 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.123292 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.226707 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.226814 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.226832 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.226860 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.226879 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.329840 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.329910 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.329923 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.329939 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.329994 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.392439 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.392494 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.392537 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:50 crc kubenswrapper[4848]: E1204 13:49:50.392637 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.392658 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:50 crc kubenswrapper[4848]: E1204 13:49:50.392795 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:50 crc kubenswrapper[4848]: E1204 13:49:50.392880 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:50 crc kubenswrapper[4848]: E1204 13:49:50.393126 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.432511 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.432548 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.432558 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.432579 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.432592 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.535295 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.535360 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.535376 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.535401 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.535419 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.638329 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.638394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.638411 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.638434 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.638453 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.741553 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.741623 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.741645 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.741668 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.741684 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.845721 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.845795 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.845812 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.845837 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.845857 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.949269 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.949339 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.949356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.949381 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:50 crc kubenswrapper[4848]: I1204 13:49:50.949399 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:50Z","lastTransitionTime":"2025-12-04T13:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.053305 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.053345 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.053355 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.053368 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.053377 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.111727 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.111818 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.111840 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.111873 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.111931 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: E1204 13:49:51.134413 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.139704 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.139755 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.139774 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.139796 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.139813 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: E1204 13:49:51.161027 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.165862 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.165909 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.165926 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.165974 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.165992 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: E1204 13:49:51.186726 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.221159 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.221253 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.221278 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.221309 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.221331 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: E1204 13:49:51.245161 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.251523 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.251597 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.251621 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.251656 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.251684 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: E1204 13:49:51.274175 4848 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c51146e5-a7ba-4456-8c4a-822438e48eb1\\\",\\\"systemUUID\\\":\\\"fca2e846-ef30-4639-ace9-fc8f55c29a70\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:51Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:51 crc kubenswrapper[4848]: E1204 13:49:51.274444 4848 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.277480 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.277526 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.277538 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.277558 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.277571 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.380614 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.380675 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.380697 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.380726 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.380747 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.482503 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.482610 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.482634 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.483087 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.483161 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.586192 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.586267 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.586289 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.586317 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.586336 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.689374 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.689432 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.689449 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.689473 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.689490 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.793032 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.793108 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.793130 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.793157 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.793175 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.896054 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.896383 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.896400 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.896425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:51 crc kubenswrapper[4848]: I1204 13:49:51.896443 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:51Z","lastTransitionTime":"2025-12-04T13:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.000317 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.000384 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.000398 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.000416 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.000429 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.103867 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.103916 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.103931 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.104003 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.104030 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.206529 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.206599 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.206617 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.206641 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.206660 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.309447 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.309515 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.309534 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.309559 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.309577 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.392843 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.392982 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:52 crc kubenswrapper[4848]: E1204 13:49:52.393087 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.393127 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.393170 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:52 crc kubenswrapper[4848]: E1204 13:49:52.393319 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:52 crc kubenswrapper[4848]: E1204 13:49:52.393444 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:52 crc kubenswrapper[4848]: E1204 13:49:52.393611 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.412259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.412333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.412359 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.412386 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.412409 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.515149 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.515236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.515270 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.515301 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.515322 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.618730 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.618816 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.618840 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.618870 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.618893 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.721373 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.721435 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.721457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.721486 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.721508 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.824089 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.824163 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.824187 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.824218 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.824240 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.927423 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.927490 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.927510 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.927538 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:52 crc kubenswrapper[4848]: I1204 13:49:52.927558 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:52Z","lastTransitionTime":"2025-12-04T13:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.029681 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.029738 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.029760 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.029788 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.029809 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.133354 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.133419 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.133457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.133488 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.133513 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.235571 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.235613 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.235623 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.235641 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.235652 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.338084 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.338135 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.338147 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.338164 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.338180 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.394650 4848 scope.go:117] "RemoveContainer" containerID="83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02" Dec 04 13:49:53 crc kubenswrapper[4848]: E1204 13:49:53.395107 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.442364 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.442441 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.442464 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.442501 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.442525 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.545871 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.545942 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.545972 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.545991 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.546002 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.649162 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.649264 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.649284 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.649308 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.649325 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.757089 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.757163 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.757198 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.757226 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.757245 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.860698 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.860919 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.860972 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.861001 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.861021 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.965207 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.965287 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.965312 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.965340 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:53 crc kubenswrapper[4848]: I1204 13:49:53.965362 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:53Z","lastTransitionTime":"2025-12-04T13:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.069145 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.069200 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.069216 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.069240 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.069257 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.172369 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.172436 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.172457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.172483 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.172502 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.275006 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.275092 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.275115 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.275144 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.275165 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.377994 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.378016 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.378024 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.378037 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.378046 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.393212 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.393356 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:54 crc kubenswrapper[4848]: E1204 13:49:54.393623 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.393662 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.393416 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:54 crc kubenswrapper[4848]: E1204 13:49:54.393832 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:54 crc kubenswrapper[4848]: E1204 13:49:54.394011 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:54 crc kubenswrapper[4848]: E1204 13:49:54.394109 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.480393 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.480459 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.480476 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.480498 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.480516 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.583194 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.583253 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.583269 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.583292 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.583311 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.685680 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.685759 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.685776 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.685799 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.685815 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.789335 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.789408 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.789425 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.789453 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.789471 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.892846 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.892898 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.892921 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.893009 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.893046 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.996743 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.996840 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.996863 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.996886 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:54 crc kubenswrapper[4848]: I1204 13:49:54.996907 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:54Z","lastTransitionTime":"2025-12-04T13:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.099929 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.100032 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.100050 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.100077 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.100095 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.203457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.203499 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.203514 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.203535 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.203549 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.307708 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.307758 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.307774 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.307800 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.307817 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.411445 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.411552 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.411568 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.411593 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.411604 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.515356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.515487 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.515514 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.515547 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.515570 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.618784 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.618848 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.618872 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.618901 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.618923 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.722250 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.722346 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.722372 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.722409 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.722435 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.826071 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.826139 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.826156 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.826183 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.826202 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.928818 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.928883 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.928900 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.928927 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:55 crc kubenswrapper[4848]: I1204 13:49:55.928987 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:55Z","lastTransitionTime":"2025-12-04T13:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.032221 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.032290 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.032310 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.032335 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.032353 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.135154 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.135185 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.135193 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.135207 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.135215 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.239258 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.239391 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.239420 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.239454 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.239492 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.341440 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.341494 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.341507 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.341525 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.341538 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.392415 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.392485 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.392497 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:56 crc kubenswrapper[4848]: E1204 13:49:56.392623 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.392812 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:56 crc kubenswrapper[4848]: E1204 13:49:56.393049 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:56 crc kubenswrapper[4848]: E1204 13:49:56.393133 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:56 crc kubenswrapper[4848]: E1204 13:49:56.393229 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.409990 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f46eae5-23b6-49e4-98ad-daaeebabe8b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e056478a6ad61eb2083bca1940601b87245ba3e8ed37c17c0a8e427c15cef5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9979f85266d21d12916f2a23a42b5b984849caee4824c86cf1b9812cec3a8ae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1846b90817b03faadeb3beedc5053b32b3b4ec4d2fddf0ea2034e32fdc46b58c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31221bb9afbbd4a8d7f4252412444e7748842b9e0ed99d4170a42620ebad8661\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:48:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:48:17Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.429780 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19ff795f7d331e08e1451dae78c20f3a5f4d0ce1f52ec723733df015b40edd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.444651 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.444696 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.444712 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.444736 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.444753 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.449451 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:34Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.466650 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5577aeec-1fac-4adc-a566-b0110dd2477b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T13:48:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d725d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T13:48:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-v4k6x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T13:49:56Z is after 2025-08-24T17:21:41Z" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.521430 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=78.521412754 podStartE2EDuration="1m18.521412754s" podCreationTimestamp="2025-12-04 13:48:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.520219154 +0000 UTC m=+100.462715692" watchObservedRunningTime="2025-12-04 13:49:56.521412754 +0000 UTC m=+100.463909292" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.550606 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.550683 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.550705 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.550733 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.550752 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.612263 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lpwtz" podStartSLOduration=74.612199201 podStartE2EDuration="1m14.612199201s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.595455136 +0000 UTC m=+100.537951664" watchObservedRunningTime="2025-12-04 13:49:56.612199201 +0000 UTC m=+100.554695759" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.637837 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.637820186 podStartE2EDuration="1m18.637820186s" podCreationTimestamp="2025-12-04 13:48:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.637709164 +0000 UTC m=+100.580205692" watchObservedRunningTime="2025-12-04 13:49:56.637820186 +0000 UTC m=+100.580316714" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.638094 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fd6s7" podStartSLOduration=74.638089503 podStartE2EDuration="1m14.638089503s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.612578071 +0000 UTC m=+100.555074639" watchObservedRunningTime="2025-12-04 13:49:56.638089503 +0000 UTC m=+100.580586031" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.654402 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.654444 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.654457 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.654474 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.654486 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.676380 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podStartSLOduration=74.67635815 podStartE2EDuration="1m14.67635815s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.653434863 +0000 UTC m=+100.595931431" watchObservedRunningTime="2025-12-04 13:49:56.67635815 +0000 UTC m=+100.618854678" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.690558 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-56gg2" podStartSLOduration=74.690539221 podStartE2EDuration="1m14.690539221s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.676568275 +0000 UTC m=+100.619064833" watchObservedRunningTime="2025-12-04 13:49:56.690539221 +0000 UTC m=+100.633035749" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.702579 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-spvsd" podStartSLOduration=74.702554319 podStartE2EDuration="1m14.702554319s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.690829359 +0000 UTC m=+100.633325907" watchObservedRunningTime="2025-12-04 13:49:56.702554319 +0000 UTC m=+100.645050867" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.703045 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-8rt9n" podStartSLOduration=74.703036561 podStartE2EDuration="1m14.703036561s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.702760984 +0000 UTC m=+100.645257512" watchObservedRunningTime="2025-12-04 13:49:56.703036561 +0000 UTC m=+100.645533099" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.727340 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=83.727322373 podStartE2EDuration="1m23.727322373s" podCreationTimestamp="2025-12-04 13:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.726550213 +0000 UTC m=+100.669046741" watchObservedRunningTime="2025-12-04 13:49:56.727322373 +0000 UTC m=+100.669818911" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.740663 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=36.740637222 podStartE2EDuration="36.740637222s" podCreationTimestamp="2025-12-04 13:49:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:49:56.740491569 +0000 UTC m=+100.682988097" watchObservedRunningTime="2025-12-04 13:49:56.740637222 +0000 UTC m=+100.683133780" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.756561 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.756609 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.756624 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.756646 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.756662 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.859356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.859400 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.859417 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.859441 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.859458 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.962046 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.962098 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.962112 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.962132 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:56 crc kubenswrapper[4848]: I1204 13:49:56.962144 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:56Z","lastTransitionTime":"2025-12-04T13:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.065137 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.065215 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.065238 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.065267 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.065289 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.168264 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.168333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.168349 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.168377 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.168402 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.271844 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.271911 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.271929 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.271986 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.272005 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.375640 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.375678 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.375687 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.375776 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.375793 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.479243 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.479286 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.479294 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.479309 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.479318 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.582591 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.582650 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.582668 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.582691 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.582707 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.686232 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.686332 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.686356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.686386 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.686409 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.789657 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.789776 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.789796 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.789823 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.789843 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.892880 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.892976 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.892999 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.893027 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.893046 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.996056 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.996118 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.996136 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.996160 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:57 crc kubenswrapper[4848]: I1204 13:49:57.996177 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:57Z","lastTransitionTime":"2025-12-04T13:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.099084 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.099147 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.099165 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.099189 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.099206 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.201770 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.201801 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.201810 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.201823 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.201832 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.303764 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.303804 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.303816 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.303833 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.303844 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.392846 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.392937 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.392974 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:49:58 crc kubenswrapper[4848]: E1204 13:49:58.393066 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.393192 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:49:58 crc kubenswrapper[4848]: E1204 13:49:58.393254 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:49:58 crc kubenswrapper[4848]: E1204 13:49:58.393394 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:49:58 crc kubenswrapper[4848]: E1204 13:49:58.393586 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.406299 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.406355 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.406376 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.406402 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.406424 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.509757 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.509828 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.509847 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.509873 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.509891 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.612338 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.612404 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.612415 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.612429 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.612439 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.716146 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.716185 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.716196 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.716211 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.716221 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.818818 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.818880 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.818899 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.818923 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.818941 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.921985 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.922042 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.922060 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.922083 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:58 crc kubenswrapper[4848]: I1204 13:49:58.922100 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:58Z","lastTransitionTime":"2025-12-04T13:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.025632 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.025694 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.025710 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.025736 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.025753 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.128183 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.128248 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.128265 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.128293 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.128310 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.231602 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.231672 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.231689 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.231715 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.231733 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.334636 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.334707 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.334723 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.334746 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.334763 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.438085 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.438152 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.438174 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.438200 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.438220 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.542017 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.542085 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.542103 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.542129 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.542148 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.645191 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.645243 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.645259 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.645284 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.645303 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.748129 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.748196 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.748213 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.748236 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.748252 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.850393 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.850456 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.850470 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.850493 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.850509 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.953301 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.953333 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.953342 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.953356 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:49:59 crc kubenswrapper[4848]: I1204 13:49:59.953365 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:49:59Z","lastTransitionTime":"2025-12-04T13:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.055864 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.055904 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.055916 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.055937 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.055984 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.158834 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.158874 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.158885 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.158900 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.158910 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.261782 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.261847 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.261871 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.261900 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.261920 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.365737 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.365811 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.365835 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.365865 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.365887 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.393515 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.393592 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.393672 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:00 crc kubenswrapper[4848]: E1204 13:50:00.393686 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:00 crc kubenswrapper[4848]: E1204 13:50:00.393788 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.393832 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:00 crc kubenswrapper[4848]: E1204 13:50:00.393906 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:00 crc kubenswrapper[4848]: E1204 13:50:00.394053 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.468739 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.468786 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.468805 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.468828 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.468845 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.571369 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.571428 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.571451 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.571477 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.571498 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.674668 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.674712 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.674724 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.674739 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.674748 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.777394 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.777461 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.777487 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.777516 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.777539 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.812093 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:00 crc kubenswrapper[4848]: E1204 13:50:00.812308 4848 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:50:00 crc kubenswrapper[4848]: E1204 13:50:00.812417 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs podName:5577aeec-1fac-4adc-a566-b0110dd2477b nodeName:}" failed. No retries permitted until 2025-12-04 13:51:04.812391508 +0000 UTC m=+168.754888076 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs") pod "network-metrics-daemon-v4k6x" (UID: "5577aeec-1fac-4adc-a566-b0110dd2477b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.880408 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.880472 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.880489 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.880513 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.880534 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.983691 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.983763 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.983780 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.983803 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:00 crc kubenswrapper[4848]: I1204 13:50:00.983823 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:00Z","lastTransitionTime":"2025-12-04T13:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.087232 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.087294 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.087316 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.087346 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.087369 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:01Z","lastTransitionTime":"2025-12-04T13:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.190472 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.190519 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.190537 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.190561 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.190577 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:01Z","lastTransitionTime":"2025-12-04T13:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.293329 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.293392 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.293411 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.293437 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.293455 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:01Z","lastTransitionTime":"2025-12-04T13:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.395249 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.395294 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.395307 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.395323 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.395336 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:01Z","lastTransitionTime":"2025-12-04T13:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.498114 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.498172 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.498188 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.498211 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.498228 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:01Z","lastTransitionTime":"2025-12-04T13:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.507143 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.507208 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.507230 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.507255 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.507280 4848 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T13:50:01Z","lastTransitionTime":"2025-12-04T13:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.573658 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75"] Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.574257 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.577382 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.577780 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.578195 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.578264 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.613184 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.613164538 podStartE2EDuration="52.613164538s" podCreationTimestamp="2025-12-04 13:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:01.613036485 +0000 UTC m=+105.555533033" watchObservedRunningTime="2025-12-04 13:50:01.613164538 +0000 UTC m=+105.555661076" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.621369 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a91eff6-bcc3-4e98-b086-30e60537b559-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.621827 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2a91eff6-bcc3-4e98-b086-30e60537b559-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.621879 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2a91eff6-bcc3-4e98-b086-30e60537b559-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.621935 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a91eff6-bcc3-4e98-b086-30e60537b559-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.622002 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2a91eff6-bcc3-4e98-b086-30e60537b559-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.722825 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a91eff6-bcc3-4e98-b086-30e60537b559-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.722886 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2a91eff6-bcc3-4e98-b086-30e60537b559-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.722917 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2a91eff6-bcc3-4e98-b086-30e60537b559-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.722985 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a91eff6-bcc3-4e98-b086-30e60537b559-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.723006 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2a91eff6-bcc3-4e98-b086-30e60537b559-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.723093 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2a91eff6-bcc3-4e98-b086-30e60537b559-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.723971 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2a91eff6-bcc3-4e98-b086-30e60537b559-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.724347 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2a91eff6-bcc3-4e98-b086-30e60537b559-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.733132 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a91eff6-bcc3-4e98-b086-30e60537b559-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.746709 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a91eff6-bcc3-4e98-b086-30e60537b559-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w8l75\" (UID: \"2a91eff6-bcc3-4e98-b086-30e60537b559\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: I1204 13:50:01.898561 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" Dec 04 13:50:01 crc kubenswrapper[4848]: W1204 13:50:01.921337 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a91eff6_bcc3_4e98_b086_30e60537b559.slice/crio-2ed44d12100f01fe738a365898bc690aff24dabc81ac7c9d851f955ba61703fb WatchSource:0}: Error finding container 2ed44d12100f01fe738a365898bc690aff24dabc81ac7c9d851f955ba61703fb: Status 404 returned error can't find the container with id 2ed44d12100f01fe738a365898bc690aff24dabc81ac7c9d851f955ba61703fb Dec 04 13:50:02 crc kubenswrapper[4848]: I1204 13:50:02.025474 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" event={"ID":"2a91eff6-bcc3-4e98-b086-30e60537b559","Type":"ContainerStarted","Data":"2ed44d12100f01fe738a365898bc690aff24dabc81ac7c9d851f955ba61703fb"} Dec 04 13:50:02 crc kubenswrapper[4848]: I1204 13:50:02.393230 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:02 crc kubenswrapper[4848]: I1204 13:50:02.393338 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:02 crc kubenswrapper[4848]: I1204 13:50:02.393244 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:02 crc kubenswrapper[4848]: E1204 13:50:02.393435 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:02 crc kubenswrapper[4848]: E1204 13:50:02.393541 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:02 crc kubenswrapper[4848]: E1204 13:50:02.393670 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:02 crc kubenswrapper[4848]: I1204 13:50:02.393691 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:02 crc kubenswrapper[4848]: E1204 13:50:02.393816 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:03 crc kubenswrapper[4848]: I1204 13:50:03.034704 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" event={"ID":"2a91eff6-bcc3-4e98-b086-30e60537b559","Type":"ContainerStarted","Data":"081eae4cbe5167a8e0ac980346b2063cc3036ca0665564502a1f7ba12bf37391"} Dec 04 13:50:03 crc kubenswrapper[4848]: I1204 13:50:03.052913 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w8l75" podStartSLOduration=81.052887673 podStartE2EDuration="1m21.052887673s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:03.051585081 +0000 UTC m=+106.994081639" watchObservedRunningTime="2025-12-04 13:50:03.052887673 +0000 UTC m=+106.995384231" Dec 04 13:50:04 crc kubenswrapper[4848]: I1204 13:50:04.392998 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:04 crc kubenswrapper[4848]: E1204 13:50:04.393103 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:04 crc kubenswrapper[4848]: I1204 13:50:04.393144 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:04 crc kubenswrapper[4848]: I1204 13:50:04.393245 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:04 crc kubenswrapper[4848]: E1204 13:50:04.393304 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:04 crc kubenswrapper[4848]: E1204 13:50:04.393409 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:04 crc kubenswrapper[4848]: I1204 13:50:04.394693 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:04 crc kubenswrapper[4848]: E1204 13:50:04.395029 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:05 crc kubenswrapper[4848]: I1204 13:50:05.394253 4848 scope.go:117] "RemoveContainer" containerID="83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02" Dec 04 13:50:05 crc kubenswrapper[4848]: E1204 13:50:05.394775 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:50:06 crc kubenswrapper[4848]: I1204 13:50:06.392584 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:06 crc kubenswrapper[4848]: I1204 13:50:06.392584 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:06 crc kubenswrapper[4848]: I1204 13:50:06.392615 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:06 crc kubenswrapper[4848]: I1204 13:50:06.392707 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:06 crc kubenswrapper[4848]: E1204 13:50:06.393836 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:06 crc kubenswrapper[4848]: E1204 13:50:06.393913 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:06 crc kubenswrapper[4848]: E1204 13:50:06.394277 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:06 crc kubenswrapper[4848]: E1204 13:50:06.394330 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:08 crc kubenswrapper[4848]: I1204 13:50:08.393000 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:08 crc kubenswrapper[4848]: I1204 13:50:08.393016 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:08 crc kubenswrapper[4848]: E1204 13:50:08.394358 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:08 crc kubenswrapper[4848]: I1204 13:50:08.393141 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:08 crc kubenswrapper[4848]: E1204 13:50:08.394896 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:08 crc kubenswrapper[4848]: E1204 13:50:08.394554 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:08 crc kubenswrapper[4848]: I1204 13:50:08.393083 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:08 crc kubenswrapper[4848]: E1204 13:50:08.395162 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:10 crc kubenswrapper[4848]: I1204 13:50:10.392762 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:10 crc kubenswrapper[4848]: I1204 13:50:10.392761 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:10 crc kubenswrapper[4848]: I1204 13:50:10.392905 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:10 crc kubenswrapper[4848]: E1204 13:50:10.393997 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:10 crc kubenswrapper[4848]: I1204 13:50:10.393022 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:10 crc kubenswrapper[4848]: E1204 13:50:10.394131 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:10 crc kubenswrapper[4848]: E1204 13:50:10.394192 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:10 crc kubenswrapper[4848]: E1204 13:50:10.394565 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:12 crc kubenswrapper[4848]: I1204 13:50:12.392868 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:12 crc kubenswrapper[4848]: E1204 13:50:12.393115 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:12 crc kubenswrapper[4848]: I1204 13:50:12.393421 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:12 crc kubenswrapper[4848]: E1204 13:50:12.393508 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:12 crc kubenswrapper[4848]: I1204 13:50:12.393743 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:12 crc kubenswrapper[4848]: E1204 13:50:12.393850 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:12 crc kubenswrapper[4848]: I1204 13:50:12.394198 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:12 crc kubenswrapper[4848]: E1204 13:50:12.394323 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:14 crc kubenswrapper[4848]: I1204 13:50:14.394119 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:14 crc kubenswrapper[4848]: I1204 13:50:14.394176 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:14 crc kubenswrapper[4848]: I1204 13:50:14.394217 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:14 crc kubenswrapper[4848]: E1204 13:50:14.394344 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:14 crc kubenswrapper[4848]: I1204 13:50:14.394613 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:14 crc kubenswrapper[4848]: E1204 13:50:14.394719 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:14 crc kubenswrapper[4848]: E1204 13:50:14.394933 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:14 crc kubenswrapper[4848]: E1204 13:50:14.395240 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:16 crc kubenswrapper[4848]: E1204 13:50:16.370791 4848 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 04 13:50:16 crc kubenswrapper[4848]: I1204 13:50:16.392489 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:16 crc kubenswrapper[4848]: I1204 13:50:16.392534 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:16 crc kubenswrapper[4848]: I1204 13:50:16.392574 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:16 crc kubenswrapper[4848]: I1204 13:50:16.392489 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:16 crc kubenswrapper[4848]: E1204 13:50:16.394708 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:16 crc kubenswrapper[4848]: E1204 13:50:16.394907 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:16 crc kubenswrapper[4848]: E1204 13:50:16.395149 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:16 crc kubenswrapper[4848]: E1204 13:50:16.395417 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:16 crc kubenswrapper[4848]: E1204 13:50:16.482221 4848 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:50:17 crc kubenswrapper[4848]: I1204 13:50:17.090306 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/1.log" Dec 04 13:50:17 crc kubenswrapper[4848]: I1204 13:50:17.090919 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/0.log" Dec 04 13:50:17 crc kubenswrapper[4848]: I1204 13:50:17.091019 4848 generic.go:334] "Generic (PLEG): container finished" podID="2b6bff84-ab72-4936-8c9f-0508ded8fdd8" containerID="5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955" exitCode=1 Dec 04 13:50:17 crc kubenswrapper[4848]: I1204 13:50:17.091061 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spvsd" event={"ID":"2b6bff84-ab72-4936-8c9f-0508ded8fdd8","Type":"ContainerDied","Data":"5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955"} Dec 04 13:50:17 crc kubenswrapper[4848]: I1204 13:50:17.091144 4848 scope.go:117] "RemoveContainer" containerID="a3f1c2337f638b10ac5fde79711307153cafe8b3ae8f659078ad4c47fb8b2a99" Dec 04 13:50:17 crc kubenswrapper[4848]: I1204 13:50:17.093017 4848 scope.go:117] "RemoveContainer" containerID="5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955" Dec 04 13:50:17 crc kubenswrapper[4848]: E1204 13:50:17.093314 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-spvsd_openshift-multus(2b6bff84-ab72-4936-8c9f-0508ded8fdd8)\"" pod="openshift-multus/multus-spvsd" podUID="2b6bff84-ab72-4936-8c9f-0508ded8fdd8" Dec 04 13:50:18 crc kubenswrapper[4848]: I1204 13:50:18.096161 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/1.log" Dec 04 13:50:18 crc kubenswrapper[4848]: I1204 13:50:18.392759 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:18 crc kubenswrapper[4848]: I1204 13:50:18.392777 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:18 crc kubenswrapper[4848]: I1204 13:50:18.393012 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:18 crc kubenswrapper[4848]: E1204 13:50:18.393027 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:18 crc kubenswrapper[4848]: I1204 13:50:18.394067 4848 scope.go:117] "RemoveContainer" containerID="83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02" Dec 04 13:50:18 crc kubenswrapper[4848]: E1204 13:50:18.394447 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-vwdkp_openshift-ovn-kubernetes(88de7a37-3579-4630-9aa7-64654e68a472)\"" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" Dec 04 13:50:18 crc kubenswrapper[4848]: I1204 13:50:18.394719 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:18 crc kubenswrapper[4848]: E1204 13:50:18.394844 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:18 crc kubenswrapper[4848]: E1204 13:50:18.395153 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:18 crc kubenswrapper[4848]: E1204 13:50:18.395298 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:20 crc kubenswrapper[4848]: I1204 13:50:20.392643 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:20 crc kubenswrapper[4848]: E1204 13:50:20.392767 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:20 crc kubenswrapper[4848]: I1204 13:50:20.392869 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:20 crc kubenswrapper[4848]: E1204 13:50:20.393063 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:20 crc kubenswrapper[4848]: I1204 13:50:20.392868 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:20 crc kubenswrapper[4848]: E1204 13:50:20.393407 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:20 crc kubenswrapper[4848]: I1204 13:50:20.394404 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:20 crc kubenswrapper[4848]: E1204 13:50:20.394673 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:21 crc kubenswrapper[4848]: E1204 13:50:21.484032 4848 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:50:22 crc kubenswrapper[4848]: I1204 13:50:22.392523 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:22 crc kubenswrapper[4848]: I1204 13:50:22.392553 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:22 crc kubenswrapper[4848]: I1204 13:50:22.392674 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:22 crc kubenswrapper[4848]: E1204 13:50:22.393622 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:22 crc kubenswrapper[4848]: E1204 13:50:22.393718 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:22 crc kubenswrapper[4848]: I1204 13:50:22.392730 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:22 crc kubenswrapper[4848]: E1204 13:50:22.393850 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:22 crc kubenswrapper[4848]: E1204 13:50:22.394049 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:24 crc kubenswrapper[4848]: I1204 13:50:24.393831 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:24 crc kubenswrapper[4848]: I1204 13:50:24.394315 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:24 crc kubenswrapper[4848]: I1204 13:50:24.394459 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:24 crc kubenswrapper[4848]: I1204 13:50:24.394483 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:24 crc kubenswrapper[4848]: E1204 13:50:24.395648 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:24 crc kubenswrapper[4848]: E1204 13:50:24.395818 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:24 crc kubenswrapper[4848]: E1204 13:50:24.396103 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:24 crc kubenswrapper[4848]: E1204 13:50:24.396167 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:26 crc kubenswrapper[4848]: I1204 13:50:26.393232 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:26 crc kubenswrapper[4848]: I1204 13:50:26.393293 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:26 crc kubenswrapper[4848]: I1204 13:50:26.393320 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:26 crc kubenswrapper[4848]: E1204 13:50:26.396412 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:26 crc kubenswrapper[4848]: I1204 13:50:26.396456 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:26 crc kubenswrapper[4848]: E1204 13:50:26.396604 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:26 crc kubenswrapper[4848]: E1204 13:50:26.396713 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:26 crc kubenswrapper[4848]: E1204 13:50:26.396766 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:26 crc kubenswrapper[4848]: E1204 13:50:26.485874 4848 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:50:28 crc kubenswrapper[4848]: I1204 13:50:28.392881 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:28 crc kubenswrapper[4848]: I1204 13:50:28.392941 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:28 crc kubenswrapper[4848]: E1204 13:50:28.393118 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:28 crc kubenswrapper[4848]: I1204 13:50:28.393186 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:28 crc kubenswrapper[4848]: I1204 13:50:28.393242 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:28 crc kubenswrapper[4848]: E1204 13:50:28.393448 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:28 crc kubenswrapper[4848]: E1204 13:50:28.393804 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:28 crc kubenswrapper[4848]: E1204 13:50:28.393938 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:30 crc kubenswrapper[4848]: I1204 13:50:30.392367 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:30 crc kubenswrapper[4848]: E1204 13:50:30.392659 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:30 crc kubenswrapper[4848]: I1204 13:50:30.392681 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:30 crc kubenswrapper[4848]: I1204 13:50:30.392732 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:30 crc kubenswrapper[4848]: I1204 13:50:30.393342 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:30 crc kubenswrapper[4848]: E1204 13:50:30.393529 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:30 crc kubenswrapper[4848]: E1204 13:50:30.393712 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:30 crc kubenswrapper[4848]: E1204 13:50:30.393855 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:30 crc kubenswrapper[4848]: I1204 13:50:30.394089 4848 scope.go:117] "RemoveContainer" containerID="83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02" Dec 04 13:50:31 crc kubenswrapper[4848]: I1204 13:50:31.146280 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/3.log" Dec 04 13:50:31 crc kubenswrapper[4848]: I1204 13:50:31.150398 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerStarted","Data":"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239"} Dec 04 13:50:31 crc kubenswrapper[4848]: I1204 13:50:31.150939 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:50:31 crc kubenswrapper[4848]: I1204 13:50:31.190442 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podStartSLOduration=109.19042147 podStartE2EDuration="1m49.19042147s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:31.188688106 +0000 UTC m=+135.131184634" watchObservedRunningTime="2025-12-04 13:50:31.19042147 +0000 UTC m=+135.132918018" Dec 04 13:50:31 crc kubenswrapper[4848]: I1204 13:50:31.442908 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-v4k6x"] Dec 04 13:50:31 crc kubenswrapper[4848]: I1204 13:50:31.443018 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:31 crc kubenswrapper[4848]: E1204 13:50:31.443100 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:31 crc kubenswrapper[4848]: E1204 13:50:31.487374 4848 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 13:50:32 crc kubenswrapper[4848]: I1204 13:50:32.393042 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:32 crc kubenswrapper[4848]: E1204 13:50:32.393550 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:32 crc kubenswrapper[4848]: I1204 13:50:32.393293 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:32 crc kubenswrapper[4848]: E1204 13:50:32.393712 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:32 crc kubenswrapper[4848]: I1204 13:50:32.393073 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:32 crc kubenswrapper[4848]: I1204 13:50:32.393768 4848 scope.go:117] "RemoveContainer" containerID="5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955" Dec 04 13:50:32 crc kubenswrapper[4848]: E1204 13:50:32.393822 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:33 crc kubenswrapper[4848]: I1204 13:50:33.162245 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/1.log" Dec 04 13:50:33 crc kubenswrapper[4848]: I1204 13:50:33.162846 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spvsd" event={"ID":"2b6bff84-ab72-4936-8c9f-0508ded8fdd8","Type":"ContainerStarted","Data":"353f581d9b32e9f053062e13bc3dc93e3a02c4028418b5a55aba4ce1fb70fc53"} Dec 04 13:50:33 crc kubenswrapper[4848]: I1204 13:50:33.392778 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:33 crc kubenswrapper[4848]: E1204 13:50:33.393681 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:34 crc kubenswrapper[4848]: I1204 13:50:34.393563 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:34 crc kubenswrapper[4848]: I1204 13:50:34.393696 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:34 crc kubenswrapper[4848]: E1204 13:50:34.393852 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:34 crc kubenswrapper[4848]: E1204 13:50:34.394085 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:34 crc kubenswrapper[4848]: I1204 13:50:34.394205 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:34 crc kubenswrapper[4848]: E1204 13:50:34.394443 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:35 crc kubenswrapper[4848]: I1204 13:50:35.393000 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:35 crc kubenswrapper[4848]: E1204 13:50:35.393192 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-v4k6x" podUID="5577aeec-1fac-4adc-a566-b0110dd2477b" Dec 04 13:50:36 crc kubenswrapper[4848]: I1204 13:50:36.393042 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:36 crc kubenswrapper[4848]: I1204 13:50:36.393080 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:36 crc kubenswrapper[4848]: I1204 13:50:36.393135 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:36 crc kubenswrapper[4848]: E1204 13:50:36.395715 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 13:50:36 crc kubenswrapper[4848]: E1204 13:50:36.395866 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 13:50:36 crc kubenswrapper[4848]: E1204 13:50:36.396043 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 13:50:37 crc kubenswrapper[4848]: I1204 13:50:37.392895 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:50:37 crc kubenswrapper[4848]: I1204 13:50:37.396142 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 13:50:37 crc kubenswrapper[4848]: I1204 13:50:37.396252 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 13:50:38 crc kubenswrapper[4848]: I1204 13:50:38.393204 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:38 crc kubenswrapper[4848]: I1204 13:50:38.393285 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:38 crc kubenswrapper[4848]: I1204 13:50:38.393465 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:38 crc kubenswrapper[4848]: I1204 13:50:38.396525 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 13:50:38 crc kubenswrapper[4848]: I1204 13:50:38.396726 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 13:50:38 crc kubenswrapper[4848]: I1204 13:50:38.398246 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 13:50:38 crc kubenswrapper[4848]: I1204 13:50:38.398301 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.221584 4848 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.263031 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q9lhv"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.263494 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.263605 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.265448 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.265873 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dmbbx"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.266158 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.266424 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.266632 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-87jmz"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.266653 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.266854 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.271359 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.271404 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.272443 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.273026 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.273351 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.274108 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.274570 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.274775 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tk6sb"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.275092 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.275185 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.275259 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.275382 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.275660 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.275864 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.276224 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2cmwc"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.275990 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.276043 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.276183 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.276749 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.276855 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.276962 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.277289 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.278735 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.279223 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.279370 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.280706 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.282171 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.286266 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.286492 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.287107 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.287459 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.287918 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.288897 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.289125 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-47fkv"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.289453 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-rrtwl"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.289648 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sd8h2"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.289904 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.290286 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-47fkv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.290488 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.293992 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.294242 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.306902 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.308413 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.331157 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.333489 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-72f2r"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.333501 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.333713 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.333909 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.333977 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.334066 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.334249 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.334424 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.334645 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.335098 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.335766 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.336336 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.336416 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.336442 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.336543 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.336848 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.337421 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.337645 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e8f7ac81-45f7-4446-a6df-78123ca7226b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338249 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/35ec6b16-4844-46cd-be61-27b9d3bcfc08-auth-proxy-config\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338318 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-policies\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338342 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338392 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6njkw\" (UniqueName: \"kubernetes.io/projected/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-kube-api-access-6njkw\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338422 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-etcd-client\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338455 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338638 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338686 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338748 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-config\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338770 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-dir\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338821 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-service-ca-bundle\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338844 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338890 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/818d6511-79e7-4869-b3cb-74ac663a31f2-audit-dir\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338923 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ec6b16-4844-46cd-be61-27b9d3bcfc08-config\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.338974 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-etcd-client\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339010 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-audit-policies\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339058 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339088 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339136 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-serving-cert\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339159 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f7ac81-45f7-4446-a6df-78123ca7226b-config\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339180 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339236 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krmxw\" (UniqueName: \"kubernetes.io/projected/35ec6b16-4844-46cd-be61-27b9d3bcfc08-kube-api-access-krmxw\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339256 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-audit\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339277 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-encryption-config\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339300 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-trusted-ca\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339321 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-config\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339338 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-serving-cert\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339359 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/35ec6b16-4844-46cd-be61-27b9d3bcfc08-machine-approver-tls\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339378 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e8f7ac81-45f7-4446-a6df-78123ca7226b-images\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339398 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg6tn\" (UniqueName: \"kubernetes.io/projected/7e21ea58-83e6-407a-a433-d78ad3ac897d-kube-api-access-bg6tn\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339418 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-node-pullsecrets\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339440 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-serving-cert\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339459 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-encryption-config\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339487 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-audit-dir\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339513 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339539 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339568 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnv8h\" (UniqueName: \"kubernetes.io/projected/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-kube-api-access-hnv8h\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339603 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339625 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-config\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339649 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e21ea58-83e6-407a-a433-d78ad3ac897d-serving-cert\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339668 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339690 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7mpb\" (UniqueName: \"kubernetes.io/projected/818d6511-79e7-4869-b3cb-74ac663a31f2-kube-api-access-m7mpb\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339711 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-image-import-ca\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339730 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339750 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7pbx\" (UniqueName: \"kubernetes.io/projected/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-kube-api-access-m7pbx\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339769 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339798 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87nx5\" (UniqueName: \"kubernetes.io/projected/e8f7ac81-45f7-4446-a6df-78123ca7226b-kube-api-access-87nx5\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339818 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339837 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-etcd-serving-ca\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.339858 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340407 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q9lhv"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340446 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340468 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340746 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341302 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340782 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340811 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341396 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341456 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341512 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341532 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341583 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341609 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341686 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341721 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341726 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340837 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341802 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341695 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341832 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340869 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340908 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340934 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341922 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.340911 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.341992 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.342087 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.342091 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.342516 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.342736 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.343112 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.343466 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.343590 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.343556 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.344336 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.344452 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.344687 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.345480 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fkdjg"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.345855 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.346179 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.347650 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.350307 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-nftbz"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.350848 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.351188 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.351413 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.351673 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xc69l"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.352055 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.352106 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.352131 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.352823 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.353208 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tk6sb"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.364580 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.365575 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.381557 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.381868 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.383465 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gb89l"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.384036 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.385017 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.385428 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.385440 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.385630 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.385687 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.385731 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.385814 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.385975 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.386054 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.386070 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.386108 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.386233 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.386346 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.386458 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.386608 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.386718 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.388228 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.388269 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vv5h2"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.388474 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.388791 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.389197 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gjvt9"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.389755 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.389209 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.389278 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.389325 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.390529 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.394431 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.397240 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.399782 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.405603 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.412473 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.412973 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.413283 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-72f2r"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.413307 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.413760 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w9422"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.414037 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7crwb"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.414248 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.414277 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.414340 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.414634 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zjnwv"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.415272 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.415562 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dmbbx"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.415581 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.416381 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.417216 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.417239 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.417337 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.417525 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.419231 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.427978 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.428030 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.429734 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.432864 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.435536 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.438090 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.438566 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2cmwc"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.438579 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.438588 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.438987 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.439314 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.439384 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-47fkv"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.439405 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rrtwl"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.439419 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-87jmz"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.439434 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.439513 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.439678 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.439834 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440562 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krmxw\" (UniqueName: \"kubernetes.io/projected/35ec6b16-4844-46cd-be61-27b9d3bcfc08-kube-api-access-krmxw\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440624 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-audit\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440646 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-encryption-config\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440665 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-trusted-ca\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440690 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-ca\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440710 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-serving-cert\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440728 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzkkp\" (UniqueName: \"kubernetes.io/projected/5bd73e67-206e-44f5-b34b-e6716102b5d8-kube-api-access-kzkkp\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440743 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440761 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmscm\" (UniqueName: \"kubernetes.io/projected/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-kube-api-access-pmscm\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440787 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-config\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440805 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-serving-cert\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440821 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/35ec6b16-4844-46cd-be61-27b9d3bcfc08-machine-approver-tls\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440837 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e8f7ac81-45f7-4446-a6df-78123ca7226b-images\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440852 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg6tn\" (UniqueName: \"kubernetes.io/projected/7e21ea58-83e6-407a-a433-d78ad3ac897d-kube-api-access-bg6tn\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440867 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-node-pullsecrets\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440883 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/870459be-ead1-4020-8f99-ba32001241e2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440900 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmtg5\" (UniqueName: \"kubernetes.io/projected/870459be-ead1-4020-8f99-ba32001241e2-kube-api-access-wmtg5\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440918 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed44f8e-efc1-4a6d-be21-e97dee631132-serving-cert\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440938 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2bfbb70-2af2-456a-97d1-de83de031cb6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440972 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-serving-cert\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.440988 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-encryption-config\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441006 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp6sd\" (UniqueName: \"kubernetes.io/projected/82ae395f-c4ee-43de-a1d6-0349c6dc640b-kube-api-access-zp6sd\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441024 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3702d88c-c997-4bf3-83d0-1d3ad69137b4-trusted-ca\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441042 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a708562-fa24-48a1-b6a8-8321f4de42c8-config\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441059 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a708562-fa24-48a1-b6a8-8321f4de42c8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441077 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441094 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-audit-dir\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441111 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d67a51fd-e796-4575-b650-05b22c52c652-config-volume\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441140 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441158 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441174 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a708562-fa24-48a1-b6a8-8321f4de42c8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441191 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjm4m\" (UniqueName: \"kubernetes.io/projected/d175633c-17b1-48b5-a1ed-430f08118b6c-kube-api-access-vjm4m\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441209 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4j4n\" (UniqueName: \"kubernetes.io/projected/9ba41edc-1966-4bca-ae21-b80cdd48ea80-kube-api-access-b4j4n\") pod \"downloads-7954f5f757-47fkv\" (UID: \"9ba41edc-1966-4bca-ae21-b80cdd48ea80\") " pod="openshift-console/downloads-7954f5f757-47fkv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441225 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-service-ca\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441254 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnv8h\" (UniqueName: \"kubernetes.io/projected/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-kube-api-access-hnv8h\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441270 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-config\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441278 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-audit\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441288 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441305 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-config\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441323 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441341 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-console-config\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441358 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2eca3cbf-3dba-4fb4-b761-4e8c5b386c12-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7nhs6\" (UID: \"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441375 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-oauth-config\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441394 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e21ea58-83e6-407a-a433-d78ad3ac897d-serving-cert\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441412 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441431 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7mpb\" (UniqueName: \"kubernetes.io/projected/818d6511-79e7-4869-b3cb-74ac663a31f2-kube-api-access-m7mpb\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441449 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-image-import-ca\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441465 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441482 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441502 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kg8v\" (UniqueName: \"kubernetes.io/projected/3702d88c-c997-4bf3-83d0-1d3ad69137b4-kube-api-access-4kg8v\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441520 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7pbx\" (UniqueName: \"kubernetes.io/projected/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-kube-api-access-m7pbx\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441538 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441539 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-config\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441554 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bd73e67-206e-44f5-b34b-e6716102b5d8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441612 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-client\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441741 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442029 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e8f7ac81-45f7-4446-a6df-78123ca7226b-images\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442270 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442300 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g978m\" (UniqueName: \"kubernetes.io/projected/2eca3cbf-3dba-4fb4-b761-4e8c5b386c12-kube-api-access-g978m\") pod \"cluster-samples-operator-665b6dd947-7nhs6\" (UID: \"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442321 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87nx5\" (UniqueName: \"kubernetes.io/projected/e8f7ac81-45f7-4446-a6df-78123ca7226b-kube-api-access-87nx5\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442338 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442357 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-client-ca\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442376 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-etcd-serving-ca\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442393 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2bfbb70-2af2-456a-97d1-de83de031cb6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442411 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-policies\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442428 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442443 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6njkw\" (UniqueName: \"kubernetes.io/projected/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-kube-api-access-6njkw\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442459 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/870459be-ead1-4020-8f99-ba32001241e2-serving-cert\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442477 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e8f7ac81-45f7-4446-a6df-78123ca7226b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442494 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/35ec6b16-4844-46cd-be61-27b9d3bcfc08-auth-proxy-config\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442511 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-config\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442525 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6db061af-a6d4-432f-a22d-82cdc394863b-serving-cert\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442540 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-trusted-ca-bundle\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442555 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442572 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-etcd-client\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442588 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442603 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ae395f-c4ee-43de-a1d6-0349c6dc640b-serving-cert\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442619 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3702d88c-c997-4bf3-83d0-1d3ad69137b4-metrics-tls\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442626 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gjvt9"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442635 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-oauth-serving-cert\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442707 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442733 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-config\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442758 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-dir\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442781 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwp9h\" (UniqueName: \"kubernetes.io/projected/aed44f8e-efc1-4a6d-be21-e97dee631132-kube-api-access-gwp9h\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442802 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442821 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/818d6511-79e7-4869-b3cb-74ac663a31f2-audit-dir\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442838 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-config\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442860 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-service-ca-bundle\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442878 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-etcd-client\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442896 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3702d88c-c997-4bf3-83d0-1d3ad69137b4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442913 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkrdz\" (UniqueName: \"kubernetes.io/projected/6db061af-a6d4-432f-a22d-82cdc394863b-kube-api-access-kkrdz\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442929 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bd73e67-206e-44f5-b34b-e6716102b5d8-srv-cert\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.442992 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ec6b16-4844-46cd-be61-27b9d3bcfc08-config\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443019 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-audit-policies\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443037 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443065 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h6m2\" (UniqueName: \"kubernetes.io/projected/d67a51fd-e796-4575-b650-05b22c52c652-kube-api-access-9h6m2\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443084 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443102 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443119 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-service-ca\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443138 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d67a51fd-e796-4575-b650-05b22c52c652-metrics-tls\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443159 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-serving-cert\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443175 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfv6x\" (UniqueName: \"kubernetes.io/projected/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-kube-api-access-hfv6x\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443192 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-client-ca\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443217 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f7ac81-45f7-4446-a6df-78123ca7226b-config\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443234 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443251 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcgfj\" (UniqueName: \"kubernetes.io/projected/c2bfbb70-2af2-456a-97d1-de83de031cb6-kube-api-access-dcgfj\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.443384 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.441058 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-node-pullsecrets\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.445008 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-image-import-ca\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.445224 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-encryption-config\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.445395 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-config\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.445432 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/818d6511-79e7-4869-b3cb-74ac663a31f2-audit-dir\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.445504 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.445512 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/35ec6b16-4844-46cd-be61-27b9d3bcfc08-machine-approver-tls\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.445540 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.445777 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ec6b16-4844-46cd-be61-27b9d3bcfc08-config\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.446301 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fkdjg"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.446370 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-serving-cert\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.447150 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-etcd-serving-ca\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.447258 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.447295 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-etcd-client\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.447377 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sd8h2"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.447672 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.447823 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-audit-policies\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.448116 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.448131 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.448151 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/35ec6b16-4844-46cd-be61-27b9d3bcfc08-auth-proxy-config\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.448640 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.448651 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f7ac81-45f7-4446-a6df-78123ca7226b-config\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449012 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449036 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-audit-dir\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449107 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-config\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449138 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-dir\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449138 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-serving-cert\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449282 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449441 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-policies\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449584 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/818d6511-79e7-4869-b3cb-74ac663a31f2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.449873 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.450013 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e21ea58-83e6-407a-a433-d78ad3ac897d-service-ca-bundle\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.450051 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.450237 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xc69l"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.450268 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w9422"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.450405 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.450623 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e21ea58-83e6-407a-a433-d78ad3ac897d-serving-cert\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.450851 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.451201 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.451670 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-serving-cert\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.451855 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.452015 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.452030 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-encryption-config\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.452592 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/818d6511-79e7-4869-b3cb-74ac663a31f2-etcd-client\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.452687 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e8f7ac81-45f7-4446-a6df-78123ca7226b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.452849 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.453909 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.454222 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.454308 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-trusted-ca\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.456075 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.456424 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-5mn79"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.457061 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-5mn79" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.457765 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ltbg8"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.458219 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.461896 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.462625 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gb89l"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.463636 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.464658 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.465678 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nftbz"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.466714 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.467729 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vv5h2"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.469003 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.469742 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-5mn79"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.471459 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.471931 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zjnwv"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.471968 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.472745 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.473746 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.474850 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.475969 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.477021 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h"] Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.485459 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.505325 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.526270 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.543849 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-client\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.543882 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bd73e67-206e-44f5-b34b-e6716102b5d8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.543906 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g978m\" (UniqueName: \"kubernetes.io/projected/2eca3cbf-3dba-4fb4-b761-4e8c5b386c12-kube-api-access-g978m\") pod \"cluster-samples-operator-665b6dd947-7nhs6\" (UID: \"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.543925 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-client-ca\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.543942 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2bfbb70-2af2-456a-97d1-de83de031cb6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.543994 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/870459be-ead1-4020-8f99-ba32001241e2-serving-cert\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544013 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-config\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544029 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6db061af-a6d4-432f-a22d-82cdc394863b-serving-cert\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544045 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-trusted-ca-bundle\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544063 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544079 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3702d88c-c997-4bf3-83d0-1d3ad69137b4-metrics-tls\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544094 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-oauth-serving-cert\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544110 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ae395f-c4ee-43de-a1d6-0349c6dc640b-serving-cert\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544127 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwp9h\" (UniqueName: \"kubernetes.io/projected/aed44f8e-efc1-4a6d-be21-e97dee631132-kube-api-access-gwp9h\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544142 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-config\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544159 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkrdz\" (UniqueName: \"kubernetes.io/projected/6db061af-a6d4-432f-a22d-82cdc394863b-kube-api-access-kkrdz\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544177 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bd73e67-206e-44f5-b34b-e6716102b5d8-srv-cert\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544201 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3702d88c-c997-4bf3-83d0-1d3ad69137b4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544226 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h6m2\" (UniqueName: \"kubernetes.io/projected/d67a51fd-e796-4575-b650-05b22c52c652-kube-api-access-9h6m2\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544242 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544259 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-service-ca\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544278 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfv6x\" (UniqueName: \"kubernetes.io/projected/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-kube-api-access-hfv6x\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544296 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d67a51fd-e796-4575-b650-05b22c52c652-metrics-tls\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544319 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcgfj\" (UniqueName: \"kubernetes.io/projected/c2bfbb70-2af2-456a-97d1-de83de031cb6-kube-api-access-dcgfj\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544335 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-client-ca\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544359 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-ca\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544374 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-serving-cert\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544390 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzkkp\" (UniqueName: \"kubernetes.io/projected/5bd73e67-206e-44f5-b34b-e6716102b5d8-kube-api-access-kzkkp\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544410 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544425 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmscm\" (UniqueName: \"kubernetes.io/projected/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-kube-api-access-pmscm\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544451 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/870459be-ead1-4020-8f99-ba32001241e2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544475 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmtg5\" (UniqueName: \"kubernetes.io/projected/870459be-ead1-4020-8f99-ba32001241e2-kube-api-access-wmtg5\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544491 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed44f8e-efc1-4a6d-be21-e97dee631132-serving-cert\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544508 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2bfbb70-2af2-456a-97d1-de83de031cb6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544525 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp6sd\" (UniqueName: \"kubernetes.io/projected/82ae395f-c4ee-43de-a1d6-0349c6dc640b-kube-api-access-zp6sd\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544541 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3702d88c-c997-4bf3-83d0-1d3ad69137b4-trusted-ca\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544559 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a708562-fa24-48a1-b6a8-8321f4de42c8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544577 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a708562-fa24-48a1-b6a8-8321f4de42c8-config\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544594 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a708562-fa24-48a1-b6a8-8321f4de42c8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544609 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjm4m\" (UniqueName: \"kubernetes.io/projected/d175633c-17b1-48b5-a1ed-430f08118b6c-kube-api-access-vjm4m\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544626 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d67a51fd-e796-4575-b650-05b22c52c652-config-volume\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544648 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544679 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4j4n\" (UniqueName: \"kubernetes.io/projected/9ba41edc-1966-4bca-ae21-b80cdd48ea80-kube-api-access-b4j4n\") pod \"downloads-7954f5f757-47fkv\" (UID: \"9ba41edc-1966-4bca-ae21-b80cdd48ea80\") " pod="openshift-console/downloads-7954f5f757-47fkv" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544696 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-service-ca\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544716 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-config\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544735 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544752 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-console-config\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544772 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2eca3cbf-3dba-4fb4-b761-4e8c5b386c12-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7nhs6\" (UID: \"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544790 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-oauth-config\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544806 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.544837 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kg8v\" (UniqueName: \"kubernetes.io/projected/3702d88c-c997-4bf3-83d0-1d3ad69137b4-kube-api-access-4kg8v\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.545896 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.545912 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-oauth-serving-cert\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.546111 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-service-ca\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.546179 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/870459be-ead1-4020-8f99-ba32001241e2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.545888 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.546859 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2bfbb70-2af2-456a-97d1-de83de031cb6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.547213 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed44f8e-efc1-4a6d-be21-e97dee631132-serving-cert\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.547300 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3702d88c-c997-4bf3-83d0-1d3ad69137b4-metrics-tls\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.547803 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-config\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.547906 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.548023 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3702d88c-c997-4bf3-83d0-1d3ad69137b4-trusted-ca\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.548472 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-client-ca\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.548662 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.549039 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-config\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.549561 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-client-ca\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.549772 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.549847 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.549849 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-console-config\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.550634 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-trusted-ca-bundle\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.550707 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-oauth-config\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.551482 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-serving-cert\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.552162 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/870459be-ead1-4020-8f99-ba32001241e2-serving-cert\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.552763 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2eca3cbf-3dba-4fb4-b761-4e8c5b386c12-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7nhs6\" (UID: \"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.563531 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ae395f-c4ee-43de-a1d6-0349c6dc640b-serving-cert\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.563855 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2bfbb70-2af2-456a-97d1-de83de031cb6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.566493 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.585923 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.606540 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.617025 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d67a51fd-e796-4575-b650-05b22c52c652-config-volume\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.625576 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.636409 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a708562-fa24-48a1-b6a8-8321f4de42c8-config\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.646428 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.665401 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.685837 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.693616 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d67a51fd-e796-4575-b650-05b22c52c652-metrics-tls\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.706584 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.710816 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a708562-fa24-48a1-b6a8-8321f4de42c8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.725470 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.728618 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-service-ca\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.746442 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.765729 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.772781 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6db061af-a6d4-432f-a22d-82cdc394863b-serving-cert\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.785331 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.799118 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-client\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.805481 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.825563 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.826548 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-config\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.845944 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.866878 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.868874 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6db061af-a6d4-432f-a22d-82cdc394863b-etcd-ca\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.887158 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.906464 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.926114 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.946490 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.954614 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bd73e67-206e-44f5-b34b-e6716102b5d8-srv-cert\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:42 crc kubenswrapper[4848]: I1204 13:50:42.986869 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.007427 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.027048 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.038731 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bd73e67-206e-44f5-b34b-e6716102b5d8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.066678 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.086598 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.106261 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.138427 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.146393 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.165340 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.185172 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.207379 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.226923 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.246138 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.266554 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.286288 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.306159 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.326292 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.346031 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.371019 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.386397 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.406565 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.424019 4848 request.go:700] Waited for 1.006505795s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.426121 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.446685 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.466087 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.486136 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.507047 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.526264 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.546271 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.566311 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.586474 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.606617 4848 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.627059 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.646898 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.667040 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.667354 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 13:50:43 crc kubenswrapper[4848]: E1204 13:50:43.667412 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:52:45.667368228 +0000 UTC m=+269.609864796 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.667502 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.667606 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.669491 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.672295 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.686985 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.706766 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.726268 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.746267 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.766600 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.769204 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.769632 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.773997 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.774576 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.790944 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.806989 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.827377 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.827532 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.830119 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.830499 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.847027 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.866761 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.886644 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.906424 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.925812 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.947357 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.965394 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 13:50:43 crc kubenswrapper[4848]: I1204 13:50:43.986000 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.006364 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.025988 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.046330 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.066313 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.089093 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.107861 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.127610 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 13:50:44 crc kubenswrapper[4848]: W1204 13:50:44.138891 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-363b0948b82024bb28489f3bf5ebd27fa62e1ccaf0981a86bd022b9d47234b51 WatchSource:0}: Error finding container 363b0948b82024bb28489f3bf5ebd27fa62e1ccaf0981a86bd022b9d47234b51: Status 404 returned error can't find the container with id 363b0948b82024bb28489f3bf5ebd27fa62e1ccaf0981a86bd022b9d47234b51 Dec 04 13:50:44 crc kubenswrapper[4848]: W1204 13:50:44.140896 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-75dc5768435ce35f0d2607b1842c4db7c8523a67520858d4c1df5b5cd9037e62 WatchSource:0}: Error finding container 75dc5768435ce35f0d2607b1842c4db7c8523a67520858d4c1df5b5cd9037e62: Status 404 returned error can't find the container with id 75dc5768435ce35f0d2607b1842c4db7c8523a67520858d4c1df5b5cd9037e62 Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.145870 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.166180 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.186331 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.205078 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"75dc5768435ce35f0d2607b1842c4db7c8523a67520858d4c1df5b5cd9037e62"} Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.206367 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"363b0948b82024bb28489f3bf5ebd27fa62e1ccaf0981a86bd022b9d47234b51"} Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.230561 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krmxw\" (UniqueName: \"kubernetes.io/projected/35ec6b16-4844-46cd-be61-27b9d3bcfc08-kube-api-access-krmxw\") pod \"machine-approver-56656f9798-5p5nm\" (UID: \"35ec6b16-4844-46cd-be61-27b9d3bcfc08\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.241686 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnv8h\" (UniqueName: \"kubernetes.io/projected/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-kube-api-access-hnv8h\") pod \"oauth-openshift-558db77b4-dmbbx\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.261741 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87nx5\" (UniqueName: \"kubernetes.io/projected/e8f7ac81-45f7-4446-a6df-78123ca7226b-kube-api-access-87nx5\") pod \"machine-api-operator-5694c8668f-tk6sb\" (UID: \"e8f7ac81-45f7-4446-a6df-78123ca7226b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.292619 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7pbx\" (UniqueName: \"kubernetes.io/projected/d8666080-7b14-4d7c-bc1c-fb24ea3f05e6-kube-api-access-m7pbx\") pod \"apiserver-76f77b778f-q9lhv\" (UID: \"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6\") " pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.314022 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6njkw\" (UniqueName: \"kubernetes.io/projected/e52ecea7-0826-4ceb-8d32-f21f4c3fd181-kube-api-access-6njkw\") pod \"console-operator-58897d9998-2cmwc\" (UID: \"e52ecea7-0826-4ceb-8d32-f21f4c3fd181\") " pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.317642 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.317730 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.346917 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg6tn\" (UniqueName: \"kubernetes.io/projected/7e21ea58-83e6-407a-a433-d78ad3ac897d-kube-api-access-bg6tn\") pod \"authentication-operator-69f744f599-87jmz\" (UID: \"7e21ea58-83e6-407a-a433-d78ad3ac897d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.348708 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7mpb\" (UniqueName: \"kubernetes.io/projected/818d6511-79e7-4869-b3cb-74ac663a31f2-kube-api-access-m7mpb\") pod \"apiserver-7bbb656c7d-ls2dc\" (UID: \"818d6511-79e7-4869-b3cb-74ac663a31f2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.349254 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.366513 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.385518 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.394472 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.409292 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.424852 4848 request.go:700] Waited for 1.966464459s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.427648 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.432111 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.447878 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.447958 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.466607 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.503258 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.508259 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g978m\" (UniqueName: \"kubernetes.io/projected/2eca3cbf-3dba-4fb4-b761-4e8c5b386c12-kube-api-access-g978m\") pod \"cluster-samples-operator-665b6dd947-7nhs6\" (UID: \"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.524300 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzkkp\" (UniqueName: \"kubernetes.io/projected/5bd73e67-206e-44f5-b34b-e6716102b5d8-kube-api-access-kzkkp\") pod \"olm-operator-6b444d44fb-zd5rb\" (UID: \"5bd73e67-206e-44f5-b34b-e6716102b5d8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.535404 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.541658 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.546128 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.549132 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h6m2\" (UniqueName: \"kubernetes.io/projected/d67a51fd-e796-4575-b650-05b22c52c652-kube-api-access-9h6m2\") pod \"dns-default-nftbz\" (UID: \"d67a51fd-e796-4575-b650-05b22c52c652\") " pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.567102 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kg8v\" (UniqueName: \"kubernetes.io/projected/3702d88c-c997-4bf3-83d0-1d3ad69137b4-kube-api-access-4kg8v\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.588811 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a708562-fa24-48a1-b6a8-8321f4de42c8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-q58qh\" (UID: \"4a708562-fa24-48a1-b6a8-8321f4de42c8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.603472 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmscm\" (UniqueName: \"kubernetes.io/projected/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-kube-api-access-pmscm\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.616734 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.625314 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjm4m\" (UniqueName: \"kubernetes.io/projected/d175633c-17b1-48b5-a1ed-430f08118b6c-kube-api-access-vjm4m\") pod \"console-f9d7485db-rrtwl\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.645730 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkrdz\" (UniqueName: \"kubernetes.io/projected/6db061af-a6d4-432f-a22d-82cdc394863b-kube-api-access-kkrdz\") pod \"etcd-operator-b45778765-xc69l\" (UID: \"6db061af-a6d4-432f-a22d-82cdc394863b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.646380 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.658963 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q9lhv"] Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.672707 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwp9h\" (UniqueName: \"kubernetes.io/projected/aed44f8e-efc1-4a6d-be21-e97dee631132-kube-api-access-gwp9h\") pod \"controller-manager-879f6c89f-sd8h2\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.679329 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.685308 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dmbbx"] Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.685631 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.691480 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.692986 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp6sd\" (UniqueName: \"kubernetes.io/projected/82ae395f-c4ee-43de-a1d6-0349c6dc640b-kube-api-access-zp6sd\") pod \"route-controller-manager-6576b87f9c-lwll8\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.704230 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmtg5\" (UniqueName: \"kubernetes.io/projected/870459be-ead1-4020-8f99-ba32001241e2-kube-api-access-wmtg5\") pod \"openshift-config-operator-7777fb866f-72f2r\" (UID: \"870459be-ead1-4020-8f99-ba32001241e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.719323 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3702d88c-c997-4bf3-83d0-1d3ad69137b4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-drpvc\" (UID: \"3702d88c-c997-4bf3-83d0-1d3ad69137b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.752398 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcgfj\" (UniqueName: \"kubernetes.io/projected/c2bfbb70-2af2-456a-97d1-de83de031cb6-kube-api-access-dcgfj\") pod \"openshift-apiserver-operator-796bbdcf4f-mhlpd\" (UID: \"c2bfbb70-2af2-456a-97d1-de83de031cb6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.779575 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2jr5f\" (UID: \"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:44 crc kubenswrapper[4848]: I1204 13:50:44.811511 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4j4n\" (UniqueName: \"kubernetes.io/projected/9ba41edc-1966-4bca-ae21-b80cdd48ea80-kube-api-access-b4j4n\") pod \"downloads-7954f5f757-47fkv\" (UID: \"9ba41edc-1966-4bca-ae21-b80cdd48ea80\") " pod="openshift-console/downloads-7954f5f757-47fkv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.210000 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9002d9c138132467ec9a6be1d5fa7cbe1f7321e4cc7e22964437889dbab2a800"} Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.381846 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.381882 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.382009 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.382173 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.382490 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.386805 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.388004 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-47fkv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.388104 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.390791 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfv6x\" (UniqueName: \"kubernetes.io/projected/c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a-kube-api-access-hfv6x\") pod \"openshift-controller-manager-operator-756b6f6bc6-q5mk6\" (UID: \"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:45 crc kubenswrapper[4848]: W1204 13:50:45.392220 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8666080_7b14_4d7c_bc1c_fb24ea3f05e6.slice/crio-d9200733e46db12579199a1f76d04e7ea92f9573d9c63a15931c5c19fb8a93a5 WatchSource:0}: Error finding container d9200733e46db12579199a1f76d04e7ea92f9573d9c63a15931c5c19fb8a93a5: Status 404 returned error can't find the container with id d9200733e46db12579199a1f76d04e7ea92f9573d9c63a15931c5c19fb8a93a5 Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.393419 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-trusted-ca\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.396248 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzq79\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-kube-api-access-kzq79\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.396298 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b739fda-b42c-49be-99dc-4ff123bb8cb7-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.396340 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-tls\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.396376 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-certificates\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.396440 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-bound-sa-token\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.396521 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b739fda-b42c-49be-99dc-4ff123bb8cb7-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.396568 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.396895 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:45.896880729 +0000 UTC m=+149.839377257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.402539 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.501615 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.501744 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.001721605 +0000 UTC m=+149.944218133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.501802 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-metrics-certs\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.501840 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f455f903-027b-453f-ba32-156b60daaf61-proxy-tls\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.501941 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac83f89-9e0a-4286-abd3-6f89798aa33f-serving-cert\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.502033 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87tcj\" (UniqueName: \"kubernetes.io/projected/74527b90-33e5-4580-bee4-e71afc8a89f7-kube-api-access-87tcj\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.502130 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.502162 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5gp5\" (UniqueName: \"kubernetes.io/projected/bac83f89-9e0a-4286-abd3-6f89798aa33f-kube-api-access-f5gp5\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.502200 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b739fda-b42c-49be-99dc-4ff123bb8cb7-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.502445 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31dd1a51-8183-408a-81d1-c374ddd71af9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.502496 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-default-certificate\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.503450 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b739fda-b42c-49be-99dc-4ff123bb8cb7-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.503578 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-tls\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.503890 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-certificates\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.503965 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca4d77ac-6c91-4789-8a01-553fc2563d65-config-volume\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504221 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnk95\" (UniqueName: \"kubernetes.io/projected/55609dec-d4b0-421b-b2c9-021978925d25-kube-api-access-wnk95\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504533 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-stats-auth\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504584 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f455f903-027b-453f-ba32-156b60daaf61-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504704 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmpcz\" (UniqueName: \"kubernetes.io/projected/4f441bea-6570-425f-9825-c230b357b0ad-kube-api-access-fmpcz\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504757 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4f441bea-6570-425f-9825-c230b357b0ad-signing-key\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504782 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-bound-sa-token\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504855 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5f6z\" (UniqueName: \"kubernetes.io/projected/20b89eae-237b-4b03-9a05-93fbfb79797f-kube-api-access-t5f6z\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504903 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g48st\" (UniqueName: \"kubernetes.io/projected/41654867-fa7f-43c7-b672-c3eb6500b16c-kube-api-access-g48st\") pod \"control-plane-machine-set-operator-78cbb6b69f-qx7kz\" (UID: \"41654867-fa7f-43c7-b672-c3eb6500b16c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.504971 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc2vh\" (UniqueName: \"kubernetes.io/projected/cc696412-ff5a-4c69-bd5e-99ba88d61524-kube-api-access-dc2vh\") pod \"migrator-59844c95c7-bc5mj\" (UID: \"cc696412-ff5a-4c69-bd5e-99ba88d61524\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.505379 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-certificates\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.505393 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6hw5\" (UniqueName: \"kubernetes.io/projected/ca4d77ac-6c91-4789-8a01-553fc2563d65-kube-api-access-k6hw5\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.505466 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-socket-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.505674 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae2f23d3-ae22-4f30-826c-c9d505031236-cert\") pod \"ingress-canary-5mn79\" (UID: \"ae2f23d3-ae22-4f30-826c-c9d505031236\") " pod="openshift-ingress-canary/ingress-canary-5mn79" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.505704 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e20d06c3-e604-4343-852b-e001b8aea7d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gb89l\" (UID: \"e20d06c3-e604-4343-852b-e001b8aea7d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.505771 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4f441bea-6570-425f-9825-c230b357b0ad-signing-cabundle\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.505798 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1091f565-05a0-4198-8abc-1855cc8fc914-apiservice-cert\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.505851 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mgc8\" (UniqueName: \"kubernetes.io/projected/31dd1a51-8183-408a-81d1-c374ddd71af9-kube-api-access-7mgc8\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.506049 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afb3097f-050f-4ab7-9549-a43ba168c59d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.506112 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b739fda-b42c-49be-99dc-4ff123bb8cb7-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.506384 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.506434 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtfqh\" (UniqueName: \"kubernetes.io/projected/712576cb-af46-49e8-babe-b06adc7a4011-kube-api-access-mtfqh\") pod \"package-server-manager-789f6589d5-rqjh6\" (UID: \"712576cb-af46-49e8-babe-b06adc7a4011\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.506461 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt649\" (UniqueName: \"kubernetes.io/projected/8ccad0f8-a334-4847-8c78-56a2d3b81507-kube-api-access-kt649\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.506570 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb3097f-050f-4ab7-9549-a43ba168c59d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.507617 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d14958ee-540e-4f28-b687-503a0d580a2e-metrics-tls\") pod \"dns-operator-744455d44c-gjvt9\" (UID: \"d14958ee-540e-4f28-b687-503a0d580a2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.507660 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-registration-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.507709 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f455f903-027b-453f-ba32-156b60daaf61-images\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.507767 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7de57da-744e-40db-bbe4-bab7c4111063-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.508445 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afb3097f-050f-4ab7-9549-a43ba168c59d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.508491 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvlvl\" (UniqueName: \"kubernetes.io/projected/d14958ee-540e-4f28-b687-503a0d580a2e-kube-api-access-cvlvl\") pod \"dns-operator-744455d44c-gjvt9\" (UID: \"d14958ee-540e-4f28-b687-503a0d580a2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.508621 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhlhm\" (UniqueName: \"kubernetes.io/projected/e20d06c3-e604-4343-852b-e001b8aea7d7-kube-api-access-qhlhm\") pod \"multus-admission-controller-857f4d67dd-gb89l\" (UID: \"e20d06c3-e604-4343-852b-e001b8aea7d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.508808 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ccad0f8-a334-4847-8c78-56a2d3b81507-service-ca-bundle\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.509250 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-trusted-ca\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.511120 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j46s\" (UniqueName: \"kubernetes.io/projected/ae2f23d3-ae22-4f30-826c-c9d505031236-kube-api-access-7j46s\") pod \"ingress-canary-5mn79\" (UID: \"ae2f23d3-ae22-4f30-826c-c9d505031236\") " pod="openshift-ingress-canary/ingress-canary-5mn79" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.511506 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1091f565-05a0-4198-8abc-1855cc8fc914-tmpfs\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.511549 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/712576cb-af46-49e8-babe-b06adc7a4011-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rqjh6\" (UID: \"712576cb-af46-49e8-babe-b06adc7a4011\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.511620 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cbgv\" (UniqueName: \"kubernetes.io/projected/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-kube-api-access-4cbgv\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512010 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-trusted-ca\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512046 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzq79\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-kube-api-access-kzq79\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512092 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8g25\" (UniqueName: \"kubernetes.io/projected/f455f903-027b-453f-ba32-156b60daaf61-kube-api-access-h8g25\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512131 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac83f89-9e0a-4286-abd3-6f89798aa33f-config\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512243 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8e86888a-b828-43f7-8cd8-a4429b43d716-node-bootstrap-token\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512280 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2mwc\" (UniqueName: \"kubernetes.io/projected/8e86888a-b828-43f7-8cd8-a4429b43d716-kube-api-access-h2mwc\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512400 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31dd1a51-8183-408a-81d1-c374ddd71af9-proxy-tls\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512627 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-mountpoint-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512695 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-plugins-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512797 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/20b89eae-237b-4b03-9a05-93fbfb79797f-srv-cert\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512819 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.512939 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca4d77ac-6c91-4789-8a01-553fc2563d65-secret-volume\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.513075 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7de57da-744e-40db-bbe4-bab7c4111063-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.513092 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8e86888a-b828-43f7-8cd8-a4429b43d716-certs\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.513110 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/41654867-fa7f-43c7-b672-c3eb6500b16c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qx7kz\" (UID: \"41654867-fa7f-43c7-b672-c3eb6500b16c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.519753 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b739fda-b42c-49be-99dc-4ff123bb8cb7-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.520425 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-tls\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.521371 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7de57da-744e-40db-bbe4-bab7c4111063-config\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.521623 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.524308 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdztv\" (UniqueName: \"kubernetes.io/projected/1091f565-05a0-4198-8abc-1855cc8fc914-kube-api-access-vdztv\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.524384 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.524409 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1091f565-05a0-4198-8abc-1855cc8fc914-webhook-cert\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.524490 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/20b89eae-237b-4b03-9a05-93fbfb79797f-profile-collector-cert\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.524511 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-csi-data-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.530453 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.030436697 +0000 UTC m=+149.972933225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.547680 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-bound-sa-token\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.565996 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzq79\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-kube-api-access-kzq79\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625493 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625734 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625760 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5gp5\" (UniqueName: \"kubernetes.io/projected/bac83f89-9e0a-4286-abd3-6f89798aa33f-kube-api-access-f5gp5\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625779 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31dd1a51-8183-408a-81d1-c374ddd71af9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625826 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-default-certificate\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625843 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnk95\" (UniqueName: \"kubernetes.io/projected/55609dec-d4b0-421b-b2c9-021978925d25-kube-api-access-wnk95\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625863 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca4d77ac-6c91-4789-8a01-553fc2563d65-config-volume\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625906 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-stats-auth\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625920 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f455f903-027b-453f-ba32-156b60daaf61-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.625935 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmpcz\" (UniqueName: \"kubernetes.io/projected/4f441bea-6570-425f-9825-c230b357b0ad-kube-api-access-fmpcz\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626680 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4f441bea-6570-425f-9825-c230b357b0ad-signing-key\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626700 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5f6z\" (UniqueName: \"kubernetes.io/projected/20b89eae-237b-4b03-9a05-93fbfb79797f-kube-api-access-t5f6z\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626718 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g48st\" (UniqueName: \"kubernetes.io/projected/41654867-fa7f-43c7-b672-c3eb6500b16c-kube-api-access-g48st\") pod \"control-plane-machine-set-operator-78cbb6b69f-qx7kz\" (UID: \"41654867-fa7f-43c7-b672-c3eb6500b16c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626735 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc2vh\" (UniqueName: \"kubernetes.io/projected/cc696412-ff5a-4c69-bd5e-99ba88d61524-kube-api-access-dc2vh\") pod \"migrator-59844c95c7-bc5mj\" (UID: \"cc696412-ff5a-4c69-bd5e-99ba88d61524\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626751 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6hw5\" (UniqueName: \"kubernetes.io/projected/ca4d77ac-6c91-4789-8a01-553fc2563d65-kube-api-access-k6hw5\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626766 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-socket-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626782 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae2f23d3-ae22-4f30-826c-c9d505031236-cert\") pod \"ingress-canary-5mn79\" (UID: \"ae2f23d3-ae22-4f30-826c-c9d505031236\") " pod="openshift-ingress-canary/ingress-canary-5mn79" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626797 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e20d06c3-e604-4343-852b-e001b8aea7d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gb89l\" (UID: \"e20d06c3-e604-4343-852b-e001b8aea7d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626812 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1091f565-05a0-4198-8abc-1855cc8fc914-apiservice-cert\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626825 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4f441bea-6570-425f-9825-c230b357b0ad-signing-cabundle\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626840 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mgc8\" (UniqueName: \"kubernetes.io/projected/31dd1a51-8183-408a-81d1-c374ddd71af9-kube-api-access-7mgc8\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626856 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afb3097f-050f-4ab7-9549-a43ba168c59d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626872 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626888 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtfqh\" (UniqueName: \"kubernetes.io/projected/712576cb-af46-49e8-babe-b06adc7a4011-kube-api-access-mtfqh\") pod \"package-server-manager-789f6589d5-rqjh6\" (UID: \"712576cb-af46-49e8-babe-b06adc7a4011\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626902 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt649\" (UniqueName: \"kubernetes.io/projected/8ccad0f8-a334-4847-8c78-56a2d3b81507-kube-api-access-kt649\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626918 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb3097f-050f-4ab7-9549-a43ba168c59d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626934 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d14958ee-540e-4f28-b687-503a0d580a2e-metrics-tls\") pod \"dns-operator-744455d44c-gjvt9\" (UID: \"d14958ee-540e-4f28-b687-503a0d580a2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626984 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-registration-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.626999 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f455f903-027b-453f-ba32-156b60daaf61-images\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627013 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7de57da-744e-40db-bbe4-bab7c4111063-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627029 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afb3097f-050f-4ab7-9549-a43ba168c59d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627045 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvlvl\" (UniqueName: \"kubernetes.io/projected/d14958ee-540e-4f28-b687-503a0d580a2e-kube-api-access-cvlvl\") pod \"dns-operator-744455d44c-gjvt9\" (UID: \"d14958ee-540e-4f28-b687-503a0d580a2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627063 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhlhm\" (UniqueName: \"kubernetes.io/projected/e20d06c3-e604-4343-852b-e001b8aea7d7-kube-api-access-qhlhm\") pod \"multus-admission-controller-857f4d67dd-gb89l\" (UID: \"e20d06c3-e604-4343-852b-e001b8aea7d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627076 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ccad0f8-a334-4847-8c78-56a2d3b81507-service-ca-bundle\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627096 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j46s\" (UniqueName: \"kubernetes.io/projected/ae2f23d3-ae22-4f30-826c-c9d505031236-kube-api-access-7j46s\") pod \"ingress-canary-5mn79\" (UID: \"ae2f23d3-ae22-4f30-826c-c9d505031236\") " pod="openshift-ingress-canary/ingress-canary-5mn79" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627112 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1091f565-05a0-4198-8abc-1855cc8fc914-tmpfs\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627128 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/712576cb-af46-49e8-babe-b06adc7a4011-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rqjh6\" (UID: \"712576cb-af46-49e8-babe-b06adc7a4011\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627146 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cbgv\" (UniqueName: \"kubernetes.io/projected/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-kube-api-access-4cbgv\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627167 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8g25\" (UniqueName: \"kubernetes.io/projected/f455f903-027b-453f-ba32-156b60daaf61-kube-api-access-h8g25\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627182 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac83f89-9e0a-4286-abd3-6f89798aa33f-config\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627198 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8e86888a-b828-43f7-8cd8-a4429b43d716-node-bootstrap-token\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627213 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2mwc\" (UniqueName: \"kubernetes.io/projected/8e86888a-b828-43f7-8cd8-a4429b43d716-kube-api-access-h2mwc\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627230 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31dd1a51-8183-408a-81d1-c374ddd71af9-proxy-tls\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627246 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-mountpoint-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627262 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-plugins-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627276 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/20b89eae-237b-4b03-9a05-93fbfb79797f-srv-cert\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627293 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.627364 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.127328946 +0000 UTC m=+150.069825474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627441 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca4d77ac-6c91-4789-8a01-553fc2563d65-secret-volume\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627475 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7de57da-744e-40db-bbe4-bab7c4111063-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627501 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8e86888a-b828-43f7-8cd8-a4429b43d716-certs\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627526 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/41654867-fa7f-43c7-b672-c3eb6500b16c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qx7kz\" (UID: \"41654867-fa7f-43c7-b672-c3eb6500b16c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627563 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7de57da-744e-40db-bbe4-bab7c4111063-config\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627598 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627658 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdztv\" (UniqueName: \"kubernetes.io/projected/1091f565-05a0-4198-8abc-1855cc8fc914-kube-api-access-vdztv\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627691 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627712 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1091f565-05a0-4198-8abc-1855cc8fc914-webhook-cert\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627748 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/20b89eae-237b-4b03-9a05-93fbfb79797f-profile-collector-cert\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627769 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-csi-data-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627791 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-metrics-certs\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627812 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f455f903-027b-453f-ba32-156b60daaf61-proxy-tls\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627835 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac83f89-9e0a-4286-abd3-6f89798aa33f-serving-cert\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.627866 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87tcj\" (UniqueName: \"kubernetes.io/projected/74527b90-33e5-4580-bee4-e71afc8a89f7-kube-api-access-87tcj\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.628028 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.628726 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31dd1a51-8183-408a-81d1-c374ddd71af9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.628891 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.629700 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.129686964 +0000 UTC m=+150.072183492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.635162 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb3097f-050f-4ab7-9549-a43ba168c59d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.636059 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca4d77ac-6c91-4789-8a01-553fc2563d65-config-volume\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.637178 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca4d77ac-6c91-4789-8a01-553fc2563d65-secret-volume\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.638571 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-default-certificate\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.639118 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8e86888a-b828-43f7-8cd8-a4429b43d716-certs\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.640641 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-registration-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.641551 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f455f903-027b-453f-ba32-156b60daaf61-images\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.642659 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac83f89-9e0a-4286-abd3-6f89798aa33f-config\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.642864 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f455f903-027b-453f-ba32-156b60daaf61-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.645548 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afb3097f-050f-4ab7-9549-a43ba168c59d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.645579 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-csi-data-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.645605 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-mountpoint-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.645639 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-plugins-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.646209 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/55609dec-d4b0-421b-b2c9-021978925d25-socket-dir\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.646278 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7de57da-744e-40db-bbe4-bab7c4111063-config\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.646318 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ccad0f8-a334-4847-8c78-56a2d3b81507-service-ca-bundle\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.646557 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1091f565-05a0-4198-8abc-1855cc8fc914-tmpfs\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.647018 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1091f565-05a0-4198-8abc-1855cc8fc914-webhook-cert\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.648231 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4f441bea-6570-425f-9825-c230b357b0ad-signing-cabundle\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.650203 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4f441bea-6570-425f-9825-c230b357b0ad-signing-key\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.651600 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae2f23d3-ae22-4f30-826c-c9d505031236-cert\") pod \"ingress-canary-5mn79\" (UID: \"ae2f23d3-ae22-4f30-826c-c9d505031236\") " pod="openshift-ingress-canary/ingress-canary-5mn79" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.654458 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.655774 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/712576cb-af46-49e8-babe-b06adc7a4011-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rqjh6\" (UID: \"712576cb-af46-49e8-babe-b06adc7a4011\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.662520 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f455f903-027b-453f-ba32-156b60daaf61-proxy-tls\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.663619 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/20b89eae-237b-4b03-9a05-93fbfb79797f-srv-cert\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.664387 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31dd1a51-8183-408a-81d1-c374ddd71af9-proxy-tls\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.664799 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.666860 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e20d06c3-e604-4343-852b-e001b8aea7d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gb89l\" (UID: \"e20d06c3-e604-4343-852b-e001b8aea7d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.669816 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/41654867-fa7f-43c7-b672-c3eb6500b16c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qx7kz\" (UID: \"41654867-fa7f-43c7-b672-c3eb6500b16c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.670241 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7de57da-744e-40db-bbe4-bab7c4111063-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.670254 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d14958ee-540e-4f28-b687-503a0d580a2e-metrics-tls\") pod \"dns-operator-744455d44c-gjvt9\" (UID: \"d14958ee-540e-4f28-b687-503a0d580a2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.672374 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87tcj\" (UniqueName: \"kubernetes.io/projected/74527b90-33e5-4580-bee4-e71afc8a89f7-kube-api-access-87tcj\") pod \"marketplace-operator-79b997595-vv5h2\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.672917 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-metrics-certs\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.673573 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1091f565-05a0-4198-8abc-1855cc8fc914-apiservice-cert\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.675908 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8e86888a-b828-43f7-8cd8-a4429b43d716-node-bootstrap-token\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.676375 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc"] Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.678643 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac83f89-9e0a-4286-abd3-6f89798aa33f-serving-cert\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.681355 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8ccad0f8-a334-4847-8c78-56a2d3b81507-stats-auth\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.687631 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5gp5\" (UniqueName: \"kubernetes.io/projected/bac83f89-9e0a-4286-abd3-6f89798aa33f-kube-api-access-f5gp5\") pod \"service-ca-operator-777779d784-vzm5h\" (UID: \"bac83f89-9e0a-4286-abd3-6f89798aa33f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.691512 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/20b89eae-237b-4b03-9a05-93fbfb79797f-profile-collector-cert\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.729232 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.729421 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.729722 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.229706841 +0000 UTC m=+150.172203369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.729855 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.730138 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.230131212 +0000 UTC m=+150.172627740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.730643 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7de57da-744e-40db-bbe4-bab7c4111063-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hd6t5\" (UID: \"e7de57da-744e-40db-bbe4-bab7c4111063\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.743128 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnk95\" (UniqueName: \"kubernetes.io/projected/55609dec-d4b0-421b-b2c9-021978925d25-kube-api-access-wnk95\") pod \"csi-hostpathplugin-zjnwv\" (UID: \"55609dec-d4b0-421b-b2c9-021978925d25\") " pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.766773 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdztv\" (UniqueName: \"kubernetes.io/projected/1091f565-05a0-4198-8abc-1855cc8fc914-kube-api-access-vdztv\") pod \"packageserver-d55dfcdfc-92hs4\" (UID: \"1091f565-05a0-4198-8abc-1855cc8fc914\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.802006 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvlvl\" (UniqueName: \"kubernetes.io/projected/d14958ee-540e-4f28-b687-503a0d580a2e-kube-api-access-cvlvl\") pod \"dns-operator-744455d44c-gjvt9\" (UID: \"d14958ee-540e-4f28-b687-503a0d580a2e\") " pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.808204 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2mwc\" (UniqueName: \"kubernetes.io/projected/8e86888a-b828-43f7-8cd8-a4429b43d716-kube-api-access-h2mwc\") pod \"machine-config-server-ltbg8\" (UID: \"8e86888a-b828-43f7-8cd8-a4429b43d716\") " pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.836432 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.837010 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.336992438 +0000 UTC m=+150.279488966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.837632 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8g25\" (UniqueName: \"kubernetes.io/projected/f455f903-027b-453f-ba32-156b60daaf61-kube-api-access-h8g25\") pod \"machine-config-operator-74547568cd-9g4lk\" (UID: \"f455f903-027b-453f-ba32-156b60daaf61\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.866288 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6"] Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.871633 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cbgv\" (UniqueName: \"kubernetes.io/projected/e79fb55d-fb58-46a3-90cc-d9bb97680fd2-kube-api-access-4cbgv\") pod \"kube-storage-version-migrator-operator-b67b599dd-dl6th\" (UID: \"e79fb55d-fb58-46a3-90cc-d9bb97680fd2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.871749 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5f6z\" (UniqueName: \"kubernetes.io/projected/20b89eae-237b-4b03-9a05-93fbfb79797f-kube-api-access-t5f6z\") pod \"catalog-operator-68c6474976-qr84s\" (UID: \"20b89eae-237b-4b03-9a05-93fbfb79797f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.898618 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmpcz\" (UniqueName: \"kubernetes.io/projected/4f441bea-6570-425f-9825-c230b357b0ad-kube-api-access-fmpcz\") pod \"service-ca-9c57cc56f-w9422\" (UID: \"4f441bea-6570-425f-9825-c230b357b0ad\") " pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.904128 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.912264 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.915449 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc2vh\" (UniqueName: \"kubernetes.io/projected/cc696412-ff5a-4c69-bd5e-99ba88d61524-kube-api-access-dc2vh\") pod \"migrator-59844c95c7-bc5mj\" (UID: \"cc696412-ff5a-4c69-bd5e-99ba88d61524\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.916167 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.928649 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g48st\" (UniqueName: \"kubernetes.io/projected/41654867-fa7f-43c7-b672-c3eb6500b16c-kube-api-access-g48st\") pod \"control-plane-machine-set-operator-78cbb6b69f-qx7kz\" (UID: \"41654867-fa7f-43c7-b672-c3eb6500b16c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.938794 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:45 crc kubenswrapper[4848]: E1204 13:50:45.939131 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.439118867 +0000 UTC m=+150.381615395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.940367 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.949514 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.955579 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6hw5\" (UniqueName: \"kubernetes.io/projected/ca4d77ac-6c91-4789-8a01-553fc2563d65-kube-api-access-k6hw5\") pod \"collect-profiles-29414265-d9rhm\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.962310 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.969453 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-w9422" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.974483 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhlhm\" (UniqueName: \"kubernetes.io/projected/e20d06c3-e604-4343-852b-e001b8aea7d7-kube-api-access-qhlhm\") pod \"multus-admission-controller-857f4d67dd-gb89l\" (UID: \"e20d06c3-e604-4343-852b-e001b8aea7d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.978046 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" Dec 04 13:50:45 crc kubenswrapper[4848]: I1204 13:50:45.987219 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.001939 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.006473 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mgc8\" (UniqueName: \"kubernetes.io/projected/31dd1a51-8183-408a-81d1-c374ddd71af9-kube-api-access-7mgc8\") pod \"machine-config-controller-84d6567774-8v4d7\" (UID: \"31dd1a51-8183-408a-81d1-c374ddd71af9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.012557 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j46s\" (UniqueName: \"kubernetes.io/projected/ae2f23d3-ae22-4f30-826c-c9d505031236-kube-api-access-7j46s\") pod \"ingress-canary-5mn79\" (UID: \"ae2f23d3-ae22-4f30-826c-c9d505031236\") " pod="openshift-ingress-canary/ingress-canary-5mn79" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.014377 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.021675 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.036327 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-5mn79" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.040124 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ltbg8" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.040354 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.040618 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.540593791 +0000 UTC m=+150.483090309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.043020 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/afb3097f-050f-4ab7-9549-a43ba168c59d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f5j9s\" (UID: \"afb3097f-050f-4ab7-9549-a43ba168c59d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.045655 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtfqh\" (UniqueName: \"kubernetes.io/projected/712576cb-af46-49e8-babe-b06adc7a4011-kube-api-access-mtfqh\") pod \"package-server-manager-789f6589d5-rqjh6\" (UID: \"712576cb-af46-49e8-babe-b06adc7a4011\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.065601 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt649\" (UniqueName: \"kubernetes.io/projected/8ccad0f8-a334-4847-8c78-56a2d3b81507-kube-api-access-kt649\") pod \"router-default-5444994796-7crwb\" (UID: \"8ccad0f8-a334-4847-8c78-56a2d3b81507\") " pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.147778 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.148310 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.648298708 +0000 UTC m=+150.590795236 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: W1204 13:50:46.172032 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e86888a_b828_43f7_8cd8_a4429b43d716.slice/crio-6aca8e90bfe7ced58ee5b60de08d9eef8ff694821c98b7ab30ba762678f10f7e WatchSource:0}: Error finding container 6aca8e90bfe7ced58ee5b60de08d9eef8ff694821c98b7ab30ba762678f10f7e: Status 404 returned error can't find the container with id 6aca8e90bfe7ced58ee5b60de08d9eef8ff694821c98b7ab30ba762678f10f7e Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.197459 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.216627 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" event={"ID":"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12","Type":"ContainerStarted","Data":"7c8198a00d1567549b9fbdacc848817fdc85e14a4b15c56d26c73a2e53111607"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.233976 4848 generic.go:334] "Generic (PLEG): container finished" podID="d8666080-7b14-4d7c-bc1c-fb24ea3f05e6" containerID="150dedfc1bab2e92d4057e8ee7e244f5bf3ca5aa6b41a02f910db3a6e739ab80" exitCode=0 Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.234098 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" event={"ID":"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6","Type":"ContainerDied","Data":"150dedfc1bab2e92d4057e8ee7e244f5bf3ca5aa6b41a02f910db3a6e739ab80"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.234190 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" event={"ID":"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6","Type":"ContainerStarted","Data":"d9200733e46db12579199a1f76d04e7ea92f9573d9c63a15931c5c19fb8a93a5"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.237310 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"34043d5f8dab885a1c7a7af9707bb9c74acddf15ed8813188164f584c6e616dc"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.239645 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" event={"ID":"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c","Type":"ContainerStarted","Data":"616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.239684 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" event={"ID":"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c","Type":"ContainerStarted","Data":"3da91df77c615951cb70d76df6e4a249d18aa20f365185064725042944bbecbc"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.239936 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.242329 4848 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dmbbx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" start-of-body= Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.242446 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" podUID="eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.242383 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ltbg8" event={"ID":"8e86888a-b828-43f7-8cd8-a4429b43d716","Type":"ContainerStarted","Data":"6aca8e90bfe7ced58ee5b60de08d9eef8ff694821c98b7ab30ba762678f10f7e"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.245353 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" event={"ID":"35ec6b16-4844-46cd-be61-27b9d3bcfc08","Type":"ContainerStarted","Data":"51f266a074c35d7cf554335be1e3101bbcb21d5799d598863c967a4f2bad0b10"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.245378 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" event={"ID":"35ec6b16-4844-46cd-be61-27b9d3bcfc08","Type":"ContainerStarted","Data":"fb4e31e9555f2b2ddfacca2559b51f151cc5ade286180781145188f0ae4bc7c6"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.247225 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d03fbd83cb3790aa13eaedd32fb209896ef7a80db46764fe7858e175286ed9bb"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.247349 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.248917 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.249196 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.749178286 +0000 UTC m=+150.691674814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.249298 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.249644 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.749637197 +0000 UTC m=+150.692133725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.249869 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"60b61ef89ca6d09dc967d9db7dbf595ec337af627d8a2c2d51aed70136013c01"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.254716 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" event={"ID":"818d6511-79e7-4869-b3cb-74ac663a31f2","Type":"ContainerStarted","Data":"3fc3840001a56c973c5658bde2bbefc85036160c4c2feae65dfdae1cd42d3821"} Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.257394 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.281393 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.294702 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.306172 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.372903 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.374349 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.874324255 +0000 UTC m=+150.816820803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.444496 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sd8h2"] Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.476393 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.476749 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:46.976733891 +0000 UTC m=+150.919230419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.522790 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-87jmz"] Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.576898 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.577243 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.07722794 +0000 UTC m=+151.019724468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.645882 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc"] Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.679398 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.679762 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.179749529 +0000 UTC m=+151.122246047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: W1204 13:50:46.709896 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e21ea58_83e6_407a_a433_d78ad3ac897d.slice/crio-de12d6e0e6feebf92f29ab12c18555696e8f29d042e0a3bf4dfc3486f9c755c4 WatchSource:0}: Error finding container de12d6e0e6feebf92f29ab12c18555696e8f29d042e0a3bf4dfc3486f9c755c4: Status 404 returned error can't find the container with id de12d6e0e6feebf92f29ab12c18555696e8f29d042e0a3bf4dfc3486f9c755c4 Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.782495 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.782897 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.282874353 +0000 UTC m=+151.225370871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.783024 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.783313 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.283299853 +0000 UTC m=+151.225796381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.884006 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.884664 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.384646174 +0000 UTC m=+151.327142702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.900717 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" podStartSLOduration=125.9006853 podStartE2EDuration="2m5.9006853s" podCreationTimestamp="2025-12-04 13:48:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:46.899243195 +0000 UTC m=+150.841739723" watchObservedRunningTime="2025-12-04 13:50:46.9006853 +0000 UTC m=+150.843181828" Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.905359 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh"] Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.928914 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2cmwc"] Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.944610 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tk6sb"] Dec 04 13:50:46 crc kubenswrapper[4848]: I1204 13:50:46.986456 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:46 crc kubenswrapper[4848]: E1204 13:50:46.986841 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.486827114 +0000 UTC m=+151.429323642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.087271 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.087424 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.587401435 +0000 UTC m=+151.529897953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.087680 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.087987 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.587976499 +0000 UTC m=+151.530473027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.191186 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.192366 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.692348623 +0000 UTC m=+151.634845151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.263838 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2cmwc" event={"ID":"e52ecea7-0826-4ceb-8d32-f21f4c3fd181","Type":"ContainerStarted","Data":"9e8eb6bdac5c4fd218c2fdd9e00588f65a9648d5d4d686aca7e817518f90912b"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.265518 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" event={"ID":"e8f7ac81-45f7-4446-a6df-78123ca7226b","Type":"ContainerStarted","Data":"7dde783b37b088854abfdf3b11043cb90e6ef70e51783e4c0fd39b729f37433f"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.266626 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ltbg8" event={"ID":"8e86888a-b828-43f7-8cd8-a4429b43d716","Type":"ContainerStarted","Data":"85a0b36de442c97cf571f018f882d582b2afd69a5977632c2ae25a1279d82d64"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.269230 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" event={"ID":"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12","Type":"ContainerStarted","Data":"8fe243f31a77709dc800937f2f425ab925bb5c97b7fbdb2141137e7e83238fef"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.269249 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" event={"ID":"2eca3cbf-3dba-4fb4-b761-4e8c5b386c12","Type":"ContainerStarted","Data":"d3c74c54c54bd224eacea5a847bc861fb2ddf2421c3993ff7d00b5a70ea94863"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.278086 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" event={"ID":"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6","Type":"ContainerStarted","Data":"5145ac7d9b7bf91a2422ca2a04957333ee7f695dc5152199dce2f4755be457d5"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.281936 4848 generic.go:334] "Generic (PLEG): container finished" podID="818d6511-79e7-4869-b3cb-74ac663a31f2" containerID="16d4aab8dae00d2ac2293da4945e7d51bfef3d334b5b2c8f7f8434fe8a2fa31b" exitCode=0 Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.282054 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" event={"ID":"818d6511-79e7-4869-b3cb-74ac663a31f2","Type":"ContainerDied","Data":"16d4aab8dae00d2ac2293da4945e7d51bfef3d334b5b2c8f7f8434fe8a2fa31b"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.284019 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" event={"ID":"35ec6b16-4844-46cd-be61-27b9d3bcfc08","Type":"ContainerStarted","Data":"dcf9bf3486cbce85f8f7f4d8ca35f538e73fdcb4fe783cbb26774b250d999df5"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.292669 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.293001 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.792989376 +0000 UTC m=+151.735485904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.297719 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" event={"ID":"3702d88c-c997-4bf3-83d0-1d3ad69137b4","Type":"ContainerStarted","Data":"9130ed8b9a6b26bd8e2aa24f0a8bcb905b61204b1dfa0e451d68ad8eac7c457d"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.297776 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" event={"ID":"3702d88c-c997-4bf3-83d0-1d3ad69137b4","Type":"ContainerStarted","Data":"ecd85a5fbde39ada4324772a15d34576b2a02f7bfd7d284fa72a78c17988b0a5"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.297786 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" event={"ID":"3702d88c-c997-4bf3-83d0-1d3ad69137b4","Type":"ContainerStarted","Data":"2e44ee8fd3c28734882fad97d03794b7af86c084a27509be7898ba8a2099a941"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.326313 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.326530 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" event={"ID":"7e21ea58-83e6-407a-a433-d78ad3ac897d","Type":"ContainerStarted","Data":"06226729ffe73e913c3449849fc9cb1ccf7f68b0c2df65d51a2c131e9fb14905"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.326559 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" event={"ID":"7e21ea58-83e6-407a-a433-d78ad3ac897d","Type":"ContainerStarted","Data":"de12d6e0e6feebf92f29ab12c18555696e8f29d042e0a3bf4dfc3486f9c755c4"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.338321 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" event={"ID":"aed44f8e-efc1-4a6d-be21-e97dee631132","Type":"ContainerStarted","Data":"17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.338362 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" event={"ID":"aed44f8e-efc1-4a6d-be21-e97dee631132","Type":"ContainerStarted","Data":"04368de60201de9764f1a502c5a4b2a51ab39c234239aba9842779c036f49157"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.338742 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.340191 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7crwb" event={"ID":"8ccad0f8-a334-4847-8c78-56a2d3b81507","Type":"ContainerStarted","Data":"4059f6aeb105cbb69bfc338b29b84dde2833386de0a4cdb918b162b6589a38c4"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.340219 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7crwb" event={"ID":"8ccad0f8-a334-4847-8c78-56a2d3b81507","Type":"ContainerStarted","Data":"89755c7cd5e330bb68695218118dd643274506a665e59869c539e8a8c3d11b04"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.342348 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" event={"ID":"4a708562-fa24-48a1-b6a8-8321f4de42c8","Type":"ContainerStarted","Data":"49710b3661100f4130d64aeda4edbb67fe24d0f71553330dea75e0d56daffaab"} Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.350233 4848 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-sd8h2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.350303 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" podUID="aed44f8e-efc1-4a6d-be21-e97dee631132" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.353328 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.356233 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.386006 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-72f2r"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.388182 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.397461 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.398406 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:47.898385816 +0000 UTC m=+151.840882344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.411486 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5p5nm" podStartSLOduration=126.41146972 podStartE2EDuration="2m6.41146972s" podCreationTimestamp="2025-12-04 13:48:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:47.409221444 +0000 UTC m=+151.351717972" watchObservedRunningTime="2025-12-04 13:50:47.41146972 +0000 UTC m=+151.353966238" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.411516 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rrtwl"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.425770 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6"] Dec 04 13:50:47 crc kubenswrapper[4848]: W1204 13:50:47.436170 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fbbc79b_1ab7_4bb8_a1a4_de0a6ff91db3.slice/crio-8f751ed912a1f7f1cb845bdc155801b7eb8c9813bde39149befc76bf869a4e5f WatchSource:0}: Error finding container 8f751ed912a1f7f1cb845bdc155801b7eb8c9813bde39149befc76bf869a4e5f: Status 404 returned error can't find the container with id 8f751ed912a1f7f1cb845bdc155801b7eb8c9813bde39149befc76bf869a4e5f Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.438662 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-47fkv"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.454335 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.460474 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xc69l"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.492066 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nftbz"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.493860 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7nhs6" podStartSLOduration=125.492894426 podStartE2EDuration="2m5.492894426s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:47.435349891 +0000 UTC m=+151.377846419" watchObservedRunningTime="2025-12-04 13:50:47.492894426 +0000 UTC m=+151.435390954" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.501545 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.503083 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.003063358 +0000 UTC m=+151.945559886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.506161 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-87jmz" podStartSLOduration=126.506118964 podStartE2EDuration="2m6.506118964s" podCreationTimestamp="2025-12-04 13:48:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:47.500415222 +0000 UTC m=+151.442911750" watchObservedRunningTime="2025-12-04 13:50:47.506118964 +0000 UTC m=+151.448615492" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.584497 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-drpvc" podStartSLOduration=125.584477075 podStartE2EDuration="2m5.584477075s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:47.546584226 +0000 UTC m=+151.489080754" watchObservedRunningTime="2025-12-04 13:50:47.584477075 +0000 UTC m=+151.526973603" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.604633 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.604932 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.104916811 +0000 UTC m=+152.047413339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.642684 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ltbg8" podStartSLOduration=5.642552613 podStartE2EDuration="5.642552613s" podCreationTimestamp="2025-12-04 13:50:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:47.62186405 +0000 UTC m=+151.564360588" watchObservedRunningTime="2025-12-04 13:50:47.642552613 +0000 UTC m=+151.585049141" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.647967 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.675342 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zjnwv"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.687088 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w9422"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.690123 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" podStartSLOduration=125.69010514 podStartE2EDuration="2m5.69010514s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:47.684640195 +0000 UTC m=+151.627136723" watchObservedRunningTime="2025-12-04 13:50:47.69010514 +0000 UTC m=+151.632601668" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.692187 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4"] Dec 04 13:50:47 crc kubenswrapper[4848]: W1204 13:50:47.705527 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca4d77ac_6c91_4789_8a01_553fc2563d65.slice/crio-cf1016bd99276cd90e2ced6a57abd92f1b6b18c50ea7bb619de65e98fb7b9d78 WatchSource:0}: Error finding container cf1016bd99276cd90e2ced6a57abd92f1b6b18c50ea7bb619de65e98fb7b9d78: Status 404 returned error can't find the container with id cf1016bd99276cd90e2ced6a57abd92f1b6b18c50ea7bb619de65e98fb7b9d78 Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.707343 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.707633 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.207621964 +0000 UTC m=+152.150118492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.728571 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gb89l"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.749985 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7crwb" podStartSLOduration=125.749971983 podStartE2EDuration="2m5.749971983s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:47.747705717 +0000 UTC m=+151.690202245" watchObservedRunningTime="2025-12-04 13:50:47.749971983 +0000 UTC m=+151.692468511" Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.751863 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.784296 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.785827 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.805979 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vv5h2"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.808597 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.808908 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.308892202 +0000 UTC m=+152.251388730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.811643 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-5mn79"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.821432 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.840010 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.844072 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gjvt9"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.844202 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.847208 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.860275 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.877005 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s"] Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.911565 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:47 crc kubenswrapper[4848]: E1204 13:50:47.912077 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.412065847 +0000 UTC m=+152.354562375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:47 crc kubenswrapper[4848]: I1204 13:50:47.965429 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6"] Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.013414 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.013764 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.513749705 +0000 UTC m=+152.456246233 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.114521 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.115021 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.615009933 +0000 UTC m=+152.557506461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.222204 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.225172 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.723410028 +0000 UTC m=+152.665906556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.261005 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.312444 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:48 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:48 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:48 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.312508 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.323990 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.324267 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.824256685 +0000 UTC m=+152.766753213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.372246 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rrtwl" event={"ID":"d175633c-17b1-48b5-a1ed-430f08118b6c","Type":"ContainerStarted","Data":"70c0021c264792e5830b8afe4b00288702b71b7e8492b397fbe99ea80c68e1ef"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.372300 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rrtwl" event={"ID":"d175633c-17b1-48b5-a1ed-430f08118b6c","Type":"ContainerStarted","Data":"aa5b70434d977b36b82fbdb39edeb8b21eddf525ffb7a790454b8618ee94319b"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.376047 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" event={"ID":"6db061af-a6d4-432f-a22d-82cdc394863b","Type":"ContainerStarted","Data":"f4e6ef2742378d8e183d4bfd808fa1e1fad32d0b3e92e08b114a7cd7aefe37b0"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.378903 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" event={"ID":"41654867-fa7f-43c7-b672-c3eb6500b16c","Type":"ContainerStarted","Data":"d0eccc9ff46d46370ac17bb5bc9789ed97ae57b739f8e0845876f13c232663bd"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.388164 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-w9422" event={"ID":"4f441bea-6570-425f-9825-c230b357b0ad","Type":"ContainerStarted","Data":"c0d31f1845b5c968123208ba7ed2c983f5dd749cc5dad03d61c76e1bedae5c1d"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.392828 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" event={"ID":"870459be-ead1-4020-8f99-ba32001241e2","Type":"ContainerStarted","Data":"bf0506c8433a7f910dc697e7dd13c4c411700645893e56c24672a6d81157fb87"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.392864 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" event={"ID":"870459be-ead1-4020-8f99-ba32001241e2","Type":"ContainerStarted","Data":"6b39a88595b30234faa67b04ab94623998c343061b30eba46be908373be411f2"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.405212 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-rrtwl" podStartSLOduration=126.4051952 podStartE2EDuration="2m6.4051952s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.400327529 +0000 UTC m=+152.342824047" watchObservedRunningTime="2025-12-04 13:50:48.4051952 +0000 UTC m=+152.347691728" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.412684 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" event={"ID":"d14958ee-540e-4f28-b687-503a0d580a2e","Type":"ContainerStarted","Data":"adc43bd66049fdf2f60e5818675a439c05320c5f6aee8935051c9eb40313bb93"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.420865 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" event={"ID":"55609dec-d4b0-421b-b2c9-021978925d25","Type":"ContainerStarted","Data":"81f0b4f0aae6095cf823a05678902eecad0e682b273593e4477f7363542b82bb"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.426001 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.426144 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.926105588 +0000 UTC m=+152.868602116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.426272 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.426588 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:48.926574239 +0000 UTC m=+152.869070777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.434083 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" event={"ID":"f455f903-027b-453f-ba32-156b60daaf61","Type":"ContainerStarted","Data":"1bb1e7d27e9e13eeccaa863bf459d2c0bc47c77ce0d80416a8e948695babf7dc"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.447113 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" event={"ID":"ca4d77ac-6c91-4789-8a01-553fc2563d65","Type":"ContainerStarted","Data":"cf1016bd99276cd90e2ced6a57abd92f1b6b18c50ea7bb619de65e98fb7b9d78"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.450224 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2cmwc" event={"ID":"e52ecea7-0826-4ceb-8d32-f21f4c3fd181","Type":"ContainerStarted","Data":"acde3eaee8ffce7ece5749dd1e98baa251c592bd08c449e3964b34ecc5b100d7"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.451243 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.463170 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" event={"ID":"e7de57da-744e-40db-bbe4-bab7c4111063","Type":"ContainerStarted","Data":"332cc4523f5947b82dde80ef28ca658b2e0e5aa58ad6a994bae138309f92c09c"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.482514 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" event={"ID":"818d6511-79e7-4869-b3cb-74ac663a31f2","Type":"ContainerStarted","Data":"3bf0209cdc2e1b122341fa13956bc6731806415fa873643f15c6369d56fa16fa"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.494295 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" event={"ID":"31dd1a51-8183-408a-81d1-c374ddd71af9","Type":"ContainerStarted","Data":"0a4bcacc6b61a1d8aeccc392bfadbe4e429dec5624931409119b56c32b854ced"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.506900 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-5mn79" event={"ID":"ae2f23d3-ae22-4f30-826c-c9d505031236","Type":"ContainerStarted","Data":"b4443c235e8685cdc2903cf8087d4dfc81358cf7a0c8ad9514c1ff2fe2635214"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.525034 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" event={"ID":"82ae395f-c4ee-43de-a1d6-0349c6dc640b","Type":"ContainerStarted","Data":"d1179f54587ec46800449bc4211a16305ff611d9ecc3eb4c092c941a3ecc21c9"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.525077 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.525086 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" event={"ID":"82ae395f-c4ee-43de-a1d6-0349c6dc640b","Type":"ContainerStarted","Data":"9bccd07947f67859f560f3f6d4906ce44127d6b6b10e29168ce6c67478af3091"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.525378 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-2cmwc" podStartSLOduration=126.525368066 podStartE2EDuration="2m6.525368066s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.491687231 +0000 UTC m=+152.434183759" watchObservedRunningTime="2025-12-04 13:50:48.525368066 +0000 UTC m=+152.467864594" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.529224 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" podStartSLOduration=126.529214 podStartE2EDuration="2m6.529214s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.525261742 +0000 UTC m=+152.467758270" watchObservedRunningTime="2025-12-04 13:50:48.529214 +0000 UTC m=+152.471710528" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.530104 4848 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-lwll8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.530170 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" podUID="82ae395f-c4ee-43de-a1d6-0349c6dc640b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.530410 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.531966 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.031937158 +0000 UTC m=+152.974433686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.542209 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" event={"ID":"afb3097f-050f-4ab7-9549-a43ba168c59d","Type":"ContainerStarted","Data":"cae0540498b6919c200faf50ad59d65256ab888f7ab4960ae7101358baa44104"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.547109 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" event={"ID":"20b89eae-237b-4b03-9a05-93fbfb79797f","Type":"ContainerStarted","Data":"2b2b8fa0f3f2a904c8ddbca3b583395b894dd27cb77fe02e9684894bdba81612"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.554153 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" event={"ID":"712576cb-af46-49e8-babe-b06adc7a4011","Type":"ContainerStarted","Data":"f91e0b09e11b74425ec4e810e25f476195ead5bec0f5cd8034ab1c5063231022"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.554832 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" podStartSLOduration=126.554815625 podStartE2EDuration="2m6.554815625s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.554382574 +0000 UTC m=+152.496879102" watchObservedRunningTime="2025-12-04 13:50:48.554815625 +0000 UTC m=+152.497312153" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.555078 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nftbz" event={"ID":"d67a51fd-e796-4575-b650-05b22c52c652","Type":"ContainerStarted","Data":"bf46165d94ce5a5f4ad86eac1426e5aaa0a1d05c7b4ff3591cb858c910c46ae0"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.556506 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" event={"ID":"e20d06c3-e604-4343-852b-e001b8aea7d7","Type":"ContainerStarted","Data":"b15e5219b4aeac785b5bd038275caa870197b8a80bdd231dfbfb83371730f5bf"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.565096 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" event={"ID":"bac83f89-9e0a-4286-abd3-6f89798aa33f","Type":"ContainerStarted","Data":"c96d8380c858f7c2c28fce2b47904557dc4083a061b9151fb5235676826a4ee6"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.580703 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-2cmwc" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.594204 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" event={"ID":"e79fb55d-fb58-46a3-90cc-d9bb97680fd2","Type":"ContainerStarted","Data":"646782e42c8659e821ba83de1b5ac6f5dce925f2fea7a30471785da4f246a54a"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.625356 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" event={"ID":"e8f7ac81-45f7-4446-a6df-78123ca7226b","Type":"ContainerStarted","Data":"1833b99ca089496ecfa888c4a6dba907f7f3e27265d2e92e22f9248a29b6b24a"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.625439 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" event={"ID":"e8f7ac81-45f7-4446-a6df-78123ca7226b","Type":"ContainerStarted","Data":"7102e35bcc12d0fe4c565ab3f6974f84a9c7d93c6c480f23b79f3f8c9be65dd6"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.636255 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.636605 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.13659011 +0000 UTC m=+153.079086638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.651218 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-tk6sb" podStartSLOduration=126.651200111 podStartE2EDuration="2m6.651200111s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.650397152 +0000 UTC m=+152.592893680" watchObservedRunningTime="2025-12-04 13:50:48.651200111 +0000 UTC m=+152.593696629" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.651922 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" event={"ID":"1091f565-05a0-4198-8abc-1855cc8fc914","Type":"ContainerStarted","Data":"417997970bd8239cd55c0a34f61e0b4eec38686f321435325cff4ee3303b74e8"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.667734 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" event={"ID":"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3","Type":"ContainerStarted","Data":"6b8d7f81edb5230769f2d72382360b6111f1f7a0c548efc339403db13397751f"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.667778 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" event={"ID":"9fbbc79b-1ab7-4bb8-a1a4-de0a6ff91db3","Type":"ContainerStarted","Data":"8f751ed912a1f7f1cb845bdc155801b7eb8c9813bde39149befc76bf869a4e5f"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.698161 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" event={"ID":"cc696412-ff5a-4c69-bd5e-99ba88d61524","Type":"ContainerStarted","Data":"5f47c1057ed15450fc8dda46daf6bd610b94089fc51b0e34c8ff1e6ccae935db"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.711006 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" event={"ID":"c2bfbb70-2af2-456a-97d1-de83de031cb6","Type":"ContainerStarted","Data":"74cab402a050843873ed2ea9f4cb9bc4aec129d8f50037b5b660e3c0673154ec"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.726367 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" event={"ID":"74527b90-33e5-4580-bee4-e71afc8a89f7","Type":"ContainerStarted","Data":"1a5d238e545c34525ddacd36fa2650e7641fb7c5b3a9f3c99bae5c70be1751b8"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.743708 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.744210 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.244186024 +0000 UTC m=+153.186682552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.744953 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.745572 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.245560059 +0000 UTC m=+153.188056587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.753553 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" event={"ID":"d8666080-7b14-4d7c-bc1c-fb24ea3f05e6","Type":"ContainerStarted","Data":"a52ef9293eef649092ec404c12309132163a7001cccc3b64fd9aba2648a15f4d"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.771269 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-47fkv" event={"ID":"9ba41edc-1966-4bca-ae21-b80cdd48ea80","Type":"ContainerStarted","Data":"5491e25ed0f272063d657b8014e09216df81afdec8cd4215abf3c5d6c3760ec1"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.771310 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-47fkv" event={"ID":"9ba41edc-1966-4bca-ae21-b80cdd48ea80","Type":"ContainerStarted","Data":"206d40024fb780fe15d8dc0aedaeb9b0806364ece851343d79954d4525aad154"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.772056 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-47fkv" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.788101 4848 patch_prober.go:28] interesting pod/downloads-7954f5f757-47fkv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.788163 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-47fkv" podUID="9ba41edc-1966-4bca-ae21-b80cdd48ea80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.790205 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2jr5f" podStartSLOduration=126.790189604 podStartE2EDuration="2m6.790189604s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.694862983 +0000 UTC m=+152.637359511" watchObservedRunningTime="2025-12-04 13:50:48.790189604 +0000 UTC m=+152.732686122" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.814721 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" podStartSLOduration=127.814706991 podStartE2EDuration="2m7.814706991s" podCreationTimestamp="2025-12-04 13:48:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.788625795 +0000 UTC m=+152.731122323" watchObservedRunningTime="2025-12-04 13:50:48.814706991 +0000 UTC m=+152.757203519" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.815237 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" event={"ID":"5bd73e67-206e-44f5-b34b-e6716102b5d8","Type":"ContainerStarted","Data":"41467fd4210c71a4144eebdd8cf28f1cbf66c236797aaaa7fc3f649f3a219178"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.815279 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" event={"ID":"5bd73e67-206e-44f5-b34b-e6716102b5d8","Type":"ContainerStarted","Data":"463680f104d2ff52e8752fc1dba558a5ab37c3d2f3696c054a49503d39824e96"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.816057 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.832225 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.844994 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-47fkv" podStartSLOduration=126.844977561 podStartE2EDuration="2m6.844977561s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.815814048 +0000 UTC m=+152.758310576" watchObservedRunningTime="2025-12-04 13:50:48.844977561 +0000 UTC m=+152.787474089" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.846487 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zd5rb" podStartSLOduration=126.846482178 podStartE2EDuration="2m6.846482178s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.844385135 +0000 UTC m=+152.786881663" watchObservedRunningTime="2025-12-04 13:50:48.846482178 +0000 UTC m=+152.788978706" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.847487 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:48 crc kubenswrapper[4848]: E1204 13:50:48.849119 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.349108623 +0000 UTC m=+153.291605141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.862387 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" event={"ID":"4a708562-fa24-48a1-b6a8-8321f4de42c8","Type":"ContainerStarted","Data":"00aa6468986a21cc41a858020ddb606f3e4595bfd8e944a73bf80497aaed6e4a"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.876840 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" event={"ID":"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a","Type":"ContainerStarted","Data":"7740d0188272c81f35ebef8a11522b033a13a1ad2a4cfda916e55fe8f90d10b5"} Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.894485 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.928689 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-q58qh" podStartSLOduration=126.928666623 podStartE2EDuration="2m6.928666623s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.916372149 +0000 UTC m=+152.858868677" watchObservedRunningTime="2025-12-04 13:50:48.928666623 +0000 UTC m=+152.871163151" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.978202 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:48 crc kubenswrapper[4848]: I1204 13:50:48.987301 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" podStartSLOduration=126.987283824 podStartE2EDuration="2m6.987283824s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:48.987279354 +0000 UTC m=+152.929775882" watchObservedRunningTime="2025-12-04 13:50:48.987283824 +0000 UTC m=+152.929780352" Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.028877 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.528858344 +0000 UTC m=+153.471354862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.088065 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.088485 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.58845705 +0000 UTC m=+153.530953578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.190679 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.190992 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.690982269 +0000 UTC m=+153.633478797 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.287681 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:49 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:49 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:49 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.287733 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.291345 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.292014 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.791999401 +0000 UTC m=+153.734495929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.392969 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.393751 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.893740651 +0000 UTC m=+153.836237179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.402278 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.402362 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.447852 4848 patch_prober.go:28] interesting pod/apiserver-76f77b778f-q9lhv container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]log ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]etcd ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/generic-apiserver-start-informers ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/max-in-flight-filter ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 04 13:50:49 crc kubenswrapper[4848]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 04 13:50:49 crc kubenswrapper[4848]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/project.openshift.io-projectcache ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/openshift.io-startinformers ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 04 13:50:49 crc kubenswrapper[4848]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 04 13:50:49 crc kubenswrapper[4848]: livez check failed Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.447904 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" podUID="d8666080-7b14-4d7c-bc1c-fb24ea3f05e6" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.498585 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.498869 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:49.998852134 +0000 UTC m=+153.941348662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.508047 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.508095 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.585170 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.600886 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.601227 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.101216359 +0000 UTC m=+154.043712877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.702497 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.702771 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.202758273 +0000 UTC m=+154.145254801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.804618 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.805228 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.305216661 +0000 UTC m=+154.247713189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.906210 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:49 crc kubenswrapper[4848]: E1204 13:50:49.906783 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.406770085 +0000 UTC m=+154.349266613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.911137 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" event={"ID":"c2bfbb70-2af2-456a-97d1-de83de031cb6","Type":"ContainerStarted","Data":"67d5fe8e0972a6f38352ba7ff6916abfda2fb226da66679d90e0bef8722a9d92"} Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.961361 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" event={"ID":"31dd1a51-8183-408a-81d1-c374ddd71af9","Type":"ContainerStarted","Data":"4e1c5c045f1e5a16c664fba554fd7d04a871ae2cdcbca39913286a29538eed25"} Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.975274 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" event={"ID":"e20d06c3-e604-4343-852b-e001b8aea7d7","Type":"ContainerStarted","Data":"c355cf0f40fb42feb2619aa64b98de1573604dcacd42af357f89634398f98974"} Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.979925 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" event={"ID":"20b89eae-237b-4b03-9a05-93fbfb79797f","Type":"ContainerStarted","Data":"99dd8f917a2e4665d8f2653bfa1ff856cbd15647920f708dff154465b9693b73"} Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.980902 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.989839 4848 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qr84s container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 04 13:50:49 crc kubenswrapper[4848]: I1204 13:50:49.989900 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" podUID="20b89eae-237b-4b03-9a05-93fbfb79797f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.007398 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" event={"ID":"41654867-fa7f-43c7-b672-c3eb6500b16c","Type":"ContainerStarted","Data":"197911b239cc115817aec798385925e6283dc95a42233d8f816b121da7b5ef0a"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.007659 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.008304 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.50828603 +0000 UTC m=+154.450782558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.020439 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mhlpd" podStartSLOduration=129.02041837 podStartE2EDuration="2m9.02041837s" podCreationTimestamp="2025-12-04 13:48:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:49.948049828 +0000 UTC m=+153.890546376" watchObservedRunningTime="2025-12-04 13:50:50.02041837 +0000 UTC m=+153.962914898" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.021531 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" podStartSLOduration=128.021524327 podStartE2EDuration="2m8.021524327s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.018904742 +0000 UTC m=+153.961401270" watchObservedRunningTime="2025-12-04 13:50:50.021524327 +0000 UTC m=+153.964020865" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.053119 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qx7kz" podStartSLOduration=128.0531042 podStartE2EDuration="2m8.0531042s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.05230875 +0000 UTC m=+153.994805278" watchObservedRunningTime="2025-12-04 13:50:50.0531042 +0000 UTC m=+153.995600728" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.087457 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nftbz" event={"ID":"d67a51fd-e796-4575-b650-05b22c52c652","Type":"ContainerStarted","Data":"d28bb6220f450ecaf84e47d0a36077f30c9a0e52e8d0a34eb04fd096565993a4"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.089666 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-5mn79" event={"ID":"ae2f23d3-ae22-4f30-826c-c9d505031236","Type":"ContainerStarted","Data":"cc272c363f21bae91cc9242b4a90fc6a94bb301f17c907bd27af1a1e77027c36"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.094948 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" event={"ID":"ca4d77ac-6c91-4789-8a01-553fc2563d65","Type":"ContainerStarted","Data":"4285e30e6e9757f3f2a3627eda467f9dda8d82ac97aca1918864f3a326c56878"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.097338 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" event={"ID":"1091f565-05a0-4198-8abc-1855cc8fc914","Type":"ContainerStarted","Data":"158d84a9c5c44edf6614d9cbb696699f9adc6eb40f8f562d824739756b21430c"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.102026 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.108420 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.109363 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.609348002 +0000 UTC m=+154.551844530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.117119 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-5mn79" podStartSLOduration=8.117102574 podStartE2EDuration="8.117102574s" podCreationTimestamp="2025-12-04 13:50:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.115545205 +0000 UTC m=+154.058041733" watchObservedRunningTime="2025-12-04 13:50:50.117102574 +0000 UTC m=+154.059599102" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.135182 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" event={"ID":"e7de57da-744e-40db-bbe4-bab7c4111063","Type":"ContainerStarted","Data":"33d3b86931e40a594fb653d4b225f502d386ac738af2e20fb060ca03fc6ac057"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.140480 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" event={"ID":"bac83f89-9e0a-4286-abd3-6f89798aa33f","Type":"ContainerStarted","Data":"daaef3830075c795851aeb2779b8f2a74f845c0c97e1d253561edc644db4e22d"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.142565 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" event={"ID":"e79fb55d-fb58-46a3-90cc-d9bb97680fd2","Type":"ContainerStarted","Data":"3d6f7418623de5c320b78d6b768a893120717e2a0b1b218aa421a5575bd6e602"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.167069 4848 generic.go:334] "Generic (PLEG): container finished" podID="870459be-ead1-4020-8f99-ba32001241e2" containerID="bf0506c8433a7f910dc697e7dd13c4c411700645893e56c24672a6d81157fb87" exitCode=0 Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.167161 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" event={"ID":"870459be-ead1-4020-8f99-ba32001241e2","Type":"ContainerDied","Data":"bf0506c8433a7f910dc697e7dd13c4c411700645893e56c24672a6d81157fb87"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.167189 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" event={"ID":"870459be-ead1-4020-8f99-ba32001241e2","Type":"ContainerStarted","Data":"b6b511a39cc3ccd0b33527363c865c4fd6924c3e7c4d9fe0459c1eac60fe02df"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.167746 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.182134 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" podStartSLOduration=129.182116095 podStartE2EDuration="2m9.182116095s" podCreationTimestamp="2025-12-04 13:48:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.149495047 +0000 UTC m=+154.091991575" watchObservedRunningTime="2025-12-04 13:50:50.182116095 +0000 UTC m=+154.124612623" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.183485 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" podStartSLOduration=128.183480818 podStartE2EDuration="2m8.183480818s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.181396106 +0000 UTC m=+154.123892634" watchObservedRunningTime="2025-12-04 13:50:50.183480818 +0000 UTC m=+154.125977336" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.201779 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vzm5h" podStartSLOduration=128.201762491 podStartE2EDuration="2m8.201762491s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.201102544 +0000 UTC m=+154.143599072" watchObservedRunningTime="2025-12-04 13:50:50.201762491 +0000 UTC m=+154.144259009" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.211228 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.213098 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.713082351 +0000 UTC m=+154.655578879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.231394 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" event={"ID":"74527b90-33e5-4580-bee4-e71afc8a89f7","Type":"ContainerStarted","Data":"f6b502d805ba0c2e60972126a16c489a7ba21a7c968d30447eee6fea5a87b08d"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.233307 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.233475 4848 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vv5h2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.233507 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" podUID="74527b90-33e5-4580-bee4-e71afc8a89f7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.258583 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q5mk6" event={"ID":"c562cc54-1b1c-4e89-aeaf-ce3c27a1a05a","Type":"ContainerStarted","Data":"b0320531ad44faf12ba9aa32a3d1aca53e6dc88c61b026143ff5fbb3303ea48f"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.278755 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:50 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:50 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:50 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.278806 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.287642 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" event={"ID":"f455f903-027b-453f-ba32-156b60daaf61","Type":"ContainerStarted","Data":"05fb0732acfe8e620fe21f0e950f052598f5205e93fcb4921cbb83b7b423e523"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.287686 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" event={"ID":"f455f903-027b-453f-ba32-156b60daaf61","Type":"ContainerStarted","Data":"4518656dbb13a7baa82db5d26d5f7c2ddacfaf663e35911a6273c507193f46c5"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.312306 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.313732 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.813717784 +0000 UTC m=+154.756214312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.371825 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" event={"ID":"cc696412-ff5a-4c69-bd5e-99ba88d61524","Type":"ContainerStarted","Data":"7922850841c1cdcc142d2e13cffe7c193e32fc55729eae2e268dde48db34e4f7"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.413735 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.415482 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:50.915465793 +0000 UTC m=+154.857962321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.437772 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" event={"ID":"712576cb-af46-49e8-babe-b06adc7a4011","Type":"ContainerStarted","Data":"7ef5ff219f20ca5aa55256f9fa37e7e2234f931e82d2f0a068b6e869e5e693aa"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.437815 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.463160 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-w9422" event={"ID":"4f441bea-6570-425f-9825-c230b357b0ad","Type":"ContainerStarted","Data":"b4a371277a45318cd899d1a20eceb133af184aac8dfe840f914eb191ff4548bb"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.490512 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" event={"ID":"6db061af-a6d4-432f-a22d-82cdc394863b","Type":"ContainerStarted","Data":"618aee0cf9bdf005f331410dd643a75fade4c226bd4f90bf130078f160af0192"} Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.490840 4848 patch_prober.go:28] interesting pod/downloads-7954f5f757-47fkv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.490880 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-47fkv" podUID="9ba41edc-1966-4bca-ae21-b80cdd48ea80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.499976 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dl6th" podStartSLOduration=128.499951346 podStartE2EDuration="2m8.499951346s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.236581113 +0000 UTC m=+154.179077641" watchObservedRunningTime="2025-12-04 13:50:50.499951346 +0000 UTC m=+154.442447874" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.500356 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" podStartSLOduration=128.500352115 podStartE2EDuration="2m8.500352115s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.498808337 +0000 UTC m=+154.441304865" watchObservedRunningTime="2025-12-04 13:50:50.500352115 +0000 UTC m=+154.442848643" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.500469 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.507241 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ls2dc" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.515425 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.516175 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.016149016 +0000 UTC m=+154.958645544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.585101 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hd6t5" podStartSLOduration=128.585083554 podStartE2EDuration="2m8.585083554s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.523377886 +0000 UTC m=+154.465874414" watchObservedRunningTime="2025-12-04 13:50:50.585083554 +0000 UTC m=+154.527580082" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.618610 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.623068 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.123052824 +0000 UTC m=+155.065549442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.690453 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" podStartSLOduration=128.690435473 podStartE2EDuration="2m8.690435473s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.58534002 +0000 UTC m=+154.527836548" watchObservedRunningTime="2025-12-04 13:50:50.690435473 +0000 UTC m=+154.632932001" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.720387 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.720770 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.220756574 +0000 UTC m=+155.163253102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.734850 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" podStartSLOduration=128.734829643 podStartE2EDuration="2m8.734829643s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.694175396 +0000 UTC m=+154.636671914" watchObservedRunningTime="2025-12-04 13:50:50.734829643 +0000 UTC m=+154.677326171" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.822935 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.823275 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.323264262 +0000 UTC m=+155.265760790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.834756 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" podStartSLOduration=128.834736946 podStartE2EDuration="2m8.834736946s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.833238899 +0000 UTC m=+154.775735427" watchObservedRunningTime="2025-12-04 13:50:50.834736946 +0000 UTC m=+154.777233474" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.835172 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-xc69l" podStartSLOduration=128.835167757 podStartE2EDuration="2m8.835167757s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:50.79169561 +0000 UTC m=+154.734192138" watchObservedRunningTime="2025-12-04 13:50:50.835167757 +0000 UTC m=+154.777664285" Dec 04 13:50:50 crc kubenswrapper[4848]: I1204 13:50:50.924502 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:50 crc kubenswrapper[4848]: E1204 13:50:50.924813 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.424798186 +0000 UTC m=+155.367294714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.026462 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.026829 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.526813683 +0000 UTC m=+155.469310211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.032850 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-w9422" podStartSLOduration=129.032836582 podStartE2EDuration="2m9.032836582s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:51.003527586 +0000 UTC m=+154.946024114" watchObservedRunningTime="2025-12-04 13:50:51.032836582 +0000 UTC m=+154.975333110" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.098053 4848 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-92hs4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.098103 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" podUID="1091f565-05a0-4198-8abc-1855cc8fc914" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.127008 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.127357 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.627339123 +0000 UTC m=+155.569835651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.228832 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.229207 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.729191355 +0000 UTC m=+155.671687883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.261675 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:51 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:51 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:51 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.261736 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.330256 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.330468 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.830438592 +0000 UTC m=+155.772935120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.330699 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.331092 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.831082018 +0000 UTC m=+155.773578546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.431936 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.432287 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:51.932268944 +0000 UTC m=+155.874765472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.498824 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" event={"ID":"d14958ee-540e-4f28-b687-503a0d580a2e","Type":"ContainerStarted","Data":"99f454eab39f7e719c940fe7be9dbb835db352754c16d0c119a9977b449f4d6e"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.498870 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" event={"ID":"d14958ee-540e-4f28-b687-503a0d580a2e","Type":"ContainerStarted","Data":"495e5206076d5e0233833c13ba79011c206c846476a3cfb4659280ba23b07be1"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.505423 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" event={"ID":"55609dec-d4b0-421b-b2c9-021978925d25","Type":"ContainerStarted","Data":"6b2041a39596080819442a8347a60ccea1a80eae5e272ca6094cc5d13ce10d2c"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.505477 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" event={"ID":"55609dec-d4b0-421b-b2c9-021978925d25","Type":"ContainerStarted","Data":"b1d828dbdba7a58f227939421a9baedccfd2fae52eab2441946ca1d72bfcc03e"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.515433 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" event={"ID":"e20d06c3-e604-4343-852b-e001b8aea7d7","Type":"ContainerStarted","Data":"f20c8f7f198d43f15f09de21e96f4ce5e2e034c64cc9ed6922533da8624ca00e"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.518690 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" event={"ID":"afb3097f-050f-4ab7-9549-a43ba168c59d","Type":"ContainerStarted","Data":"14a0589653377195b170bfdebd908c8490de08e8cb6b31b968bf719df24eb319"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.522672 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bc5mj" event={"ID":"cc696412-ff5a-4c69-bd5e-99ba88d61524","Type":"ContainerStarted","Data":"36827eb7fb94d43240660dbfa7edfcb5c0bd4b7c1a5439e70c516bcb2f0e602a"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.524934 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9g4lk" podStartSLOduration=129.524922149 podStartE2EDuration="2m9.524922149s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:51.089077735 +0000 UTC m=+155.031574263" watchObservedRunningTime="2025-12-04 13:50:51.524922149 +0000 UTC m=+155.467418677" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.530533 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" event={"ID":"712576cb-af46-49e8-babe-b06adc7a4011","Type":"ContainerStarted","Data":"dfe23f655619fd06a0f355a306c06efc759b284828b7d66feb91dc38aead1dc8"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.533877 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" event={"ID":"31dd1a51-8183-408a-81d1-c374ddd71af9","Type":"ContainerStarted","Data":"25d485955a4eee0655f879ab5bc3d32df85f7695894c419b8ee37cf7f34461cf"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.535838 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.536304 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.036289 +0000 UTC m=+155.978785528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.541366 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-gb89l" podStartSLOduration=129.541353076 podStartE2EDuration="2m9.541353076s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:51.537418198 +0000 UTC m=+155.479914726" watchObservedRunningTime="2025-12-04 13:50:51.541353076 +0000 UTC m=+155.483849604" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.541738 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-gjvt9" podStartSLOduration=129.541733345 podStartE2EDuration="2m9.541733345s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:51.525191465 +0000 UTC m=+155.467687993" watchObservedRunningTime="2025-12-04 13:50:51.541733345 +0000 UTC m=+155.484229873" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.550733 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nftbz" event={"ID":"d67a51fd-e796-4575-b650-05b22c52c652","Type":"ContainerStarted","Data":"9639d9f935ff154f2f04b1e5a2a27d3e7bb63431f916798f89cc39ddde4b36e8"} Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.550775 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-nftbz" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.553494 4848 patch_prober.go:28] interesting pod/downloads-7954f5f757-47fkv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.553529 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-47fkv" podUID="9ba41edc-1966-4bca-ae21-b80cdd48ea80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.556586 4848 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vv5h2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.556658 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" podUID="74527b90-33e5-4580-bee4-e71afc8a89f7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.566287 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f5j9s" podStartSLOduration=129.566270522 podStartE2EDuration="2m9.566270522s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:51.565785781 +0000 UTC m=+155.508282309" watchObservedRunningTime="2025-12-04 13:50:51.566270522 +0000 UTC m=+155.508767050" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.567294 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qr84s" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.595094 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8v4d7" podStartSLOduration=129.595078406 podStartE2EDuration="2m9.595078406s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:51.594311837 +0000 UTC m=+155.536808365" watchObservedRunningTime="2025-12-04 13:50:51.595078406 +0000 UTC m=+155.537574934" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.599094 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-92hs4" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.643467 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.644813 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.144792537 +0000 UTC m=+156.087289055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.672318 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-nftbz" podStartSLOduration=9.672295448 podStartE2EDuration="9.672295448s" podCreationTimestamp="2025-12-04 13:50:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:51.62713862 +0000 UTC m=+155.569635148" watchObservedRunningTime="2025-12-04 13:50:51.672295448 +0000 UTC m=+155.614791976" Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.747817 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.748155 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.248143096 +0000 UTC m=+156.190639624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.849037 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.849220 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.349189939 +0000 UTC m=+156.291686477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.849458 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.849817 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.349801364 +0000 UTC m=+156.292297882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:51 crc kubenswrapper[4848]: I1204 13:50:51.950705 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:51 crc kubenswrapper[4848]: E1204 13:50:51.951051 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.451033211 +0000 UTC m=+156.393529739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.002897 4848 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.007691 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nr5qh"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.008838 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.012279 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.026333 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nr5qh"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.052790 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.053191 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.553173891 +0000 UTC m=+156.495670419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.153573 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.153908 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-utilities\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.153931 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgk26\" (UniqueName: \"kubernetes.io/projected/f3d120a8-d58e-45b9-be60-04be599332b3-kube-api-access-fgk26\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.154001 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-catalog-content\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.154122 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.65410758 +0000 UTC m=+156.596604108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.205644 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sjljj"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.206697 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.209544 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.232344 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjljj"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.255245 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-utilities\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.255285 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgk26\" (UniqueName: \"kubernetes.io/projected/f3d120a8-d58e-45b9-be60-04be599332b3-kube-api-access-fgk26\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.255328 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-catalog-content\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.255394 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.255673 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.755660875 +0000 UTC m=+156.698157403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.256131 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-utilities\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.256554 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-catalog-content\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.261264 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:52 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:52 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:52 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.261291 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.280345 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgk26\" (UniqueName: \"kubernetes.io/projected/f3d120a8-d58e-45b9-be60-04be599332b3-kube-api-access-fgk26\") pod \"community-operators-nr5qh\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.327039 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.357002 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.357151 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.857135488 +0000 UTC m=+156.799632016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.357176 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-catalog-content\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.357200 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgbzl\" (UniqueName: \"kubernetes.io/projected/e75ab632-da17-4354-951c-476ad2e46835-kube-api-access-lgbzl\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.357247 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-utilities\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.357311 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.357543 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.857535209 +0000 UTC m=+156.800031737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.426014 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c2dzv"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.427038 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c2dzv"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.427126 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.458097 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.458189 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.9581742 +0000 UTC m=+156.900670728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.458355 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.458388 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-catalog-content\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.458407 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgbzl\" (UniqueName: \"kubernetes.io/projected/e75ab632-da17-4354-951c-476ad2e46835-kube-api-access-lgbzl\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.458448 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-utilities\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.459006 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-utilities\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.459213 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:52.959205576 +0000 UTC m=+156.901702104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.459522 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-catalog-content\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.481125 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgbzl\" (UniqueName: \"kubernetes.io/projected/e75ab632-da17-4354-951c-476ad2e46835-kube-api-access-lgbzl\") pod \"certified-operators-sjljj\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.517602 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nr5qh"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.558991 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.559110 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 13:50:53.059092719 +0000 UTC m=+157.001589247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.559396 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfblq\" (UniqueName: \"kubernetes.io/projected/00f2c84d-1e46-4563-9c78-6657bb601db3-kube-api-access-wfblq\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.559453 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-catalog-content\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.559508 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.559629 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-utilities\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: E1204 13:50:52.559778 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 13:50:53.059768226 +0000 UTC m=+157.002264754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fkdjg" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.560927 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" event={"ID":"55609dec-d4b0-421b-b2c9-021978925d25","Type":"ContainerStarted","Data":"ee9cd7f0c8e74763fe14145df4e9e10386f77b5f44bb3b536c82a187775fe23d"} Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.560998 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" event={"ID":"55609dec-d4b0-421b-b2c9-021978925d25","Type":"ContainerStarted","Data":"7f241553ab77325cb8564ab0855f3b286c3e57ed52d7e28204da5a4e1f686bf1"} Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.565267 4848 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-04T13:50:52.002932996Z","Handler":null,"Name":""} Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.565871 4848 generic.go:334] "Generic (PLEG): container finished" podID="ca4d77ac-6c91-4789-8a01-553fc2563d65" containerID="4285e30e6e9757f3f2a3627eda467f9dda8d82ac97aca1918864f3a326c56878" exitCode=0 Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.565919 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" event={"ID":"ca4d77ac-6c91-4789-8a01-553fc2563d65","Type":"ContainerDied","Data":"4285e30e6e9757f3f2a3627eda467f9dda8d82ac97aca1918864f3a326c56878"} Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.568510 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5qh" event={"ID":"f3d120a8-d58e-45b9-be60-04be599332b3","Type":"ContainerStarted","Data":"e7c3d6c69e28d6584798e9212d42a275bd0cc7c38744a4434df2924e743ef96f"} Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.573224 4848 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.573247 4848 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.576391 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.587527 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.595393 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" podStartSLOduration=10.595370508 podStartE2EDuration="10.595370508s" podCreationTimestamp="2025-12-04 13:50:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:52.582209692 +0000 UTC m=+156.524706230" watchObservedRunningTime="2025-12-04 13:50:52.595370508 +0000 UTC m=+156.537867036" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.609993 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.627233 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kp8lv"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.628416 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.630912 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kp8lv"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.663104 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.663931 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfblq\" (UniqueName: \"kubernetes.io/projected/00f2c84d-1e46-4563-9c78-6657bb601db3-kube-api-access-wfblq\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.663997 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-catalog-content\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.665683 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-catalog-content\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.665699 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-utilities\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.666063 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-utilities\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.671052 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.685865 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfblq\" (UniqueName: \"kubernetes.io/projected/00f2c84d-1e46-4563-9c78-6657bb601db3-kube-api-access-wfblq\") pod \"community-operators-c2dzv\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.742612 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.767556 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-catalog-content\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.767598 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkqg2\" (UniqueName: \"kubernetes.io/projected/f98ec77d-021a-43cd-b8e9-12738f4921f1-kube-api-access-mkqg2\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.767650 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-utilities\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.767687 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.772135 4848 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.772171 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.796788 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fkdjg\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.839674 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjljj"] Dec 04 13:50:52 crc kubenswrapper[4848]: W1204 13:50:52.846126 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode75ab632_da17_4354_951c_476ad2e46835.slice/crio-215f327d5a8796c31cc9dc05de5f1ed6087d8a66b39ffe77470f075e80b3b2d4 WatchSource:0}: Error finding container 215f327d5a8796c31cc9dc05de5f1ed6087d8a66b39ffe77470f075e80b3b2d4: Status 404 returned error can't find the container with id 215f327d5a8796c31cc9dc05de5f1ed6087d8a66b39ffe77470f075e80b3b2d4 Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.868724 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-catalog-content\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.868768 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkqg2\" (UniqueName: \"kubernetes.io/projected/f98ec77d-021a-43cd-b8e9-12738f4921f1-kube-api-access-mkqg2\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.868818 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-utilities\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.869249 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-utilities\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.869332 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-catalog-content\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.887688 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkqg2\" (UniqueName: \"kubernetes.io/projected/f98ec77d-021a-43cd-b8e9-12738f4921f1-kube-api-access-mkqg2\") pod \"certified-operators-kp8lv\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.909634 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c2dzv"] Dec 04 13:50:52 crc kubenswrapper[4848]: I1204 13:50:52.971326 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.041154 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.191749 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kp8lv"] Dec 04 13:50:53 crc kubenswrapper[4848]: W1204 13:50:53.201950 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf98ec77d_021a_43cd_b8e9_12738f4921f1.slice/crio-aedbbf795cced2e3ea02cd5698326139e2168d2be173963a6e5f406c3975baf1 WatchSource:0}: Error finding container aedbbf795cced2e3ea02cd5698326139e2168d2be173963a6e5f406c3975baf1: Status 404 returned error can't find the container with id aedbbf795cced2e3ea02cd5698326139e2168d2be173963a6e5f406c3975baf1 Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.242167 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fkdjg"] Dec 04 13:50:53 crc kubenswrapper[4848]: W1204 13:50:53.250228 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b739fda_b42c_49be_99dc_4ff123bb8cb7.slice/crio-8fabbc7a3fdb48c79c566be5061ab4219e8fc8fffa3dee3dd2a337aaeafa431b WatchSource:0}: Error finding container 8fabbc7a3fdb48c79c566be5061ab4219e8fc8fffa3dee3dd2a337aaeafa431b: Status 404 returned error can't find the container with id 8fabbc7a3fdb48c79c566be5061ab4219e8fc8fffa3dee3dd2a337aaeafa431b Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.261198 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:53 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:53 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:53 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.261285 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.575099 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" event={"ID":"9b739fda-b42c-49be-99dc-4ff123bb8cb7","Type":"ContainerStarted","Data":"8fabbc7a3fdb48c79c566be5061ab4219e8fc8fffa3dee3dd2a337aaeafa431b"} Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.576298 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp8lv" event={"ID":"f98ec77d-021a-43cd-b8e9-12738f4921f1","Type":"ContainerStarted","Data":"aedbbf795cced2e3ea02cd5698326139e2168d2be173963a6e5f406c3975baf1"} Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.577769 4848 generic.go:334] "Generic (PLEG): container finished" podID="f3d120a8-d58e-45b9-be60-04be599332b3" containerID="760ee0d53304a79ced9adbeed9c64f4c5f56137c58551b302a474e0402afed15" exitCode=0 Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.577854 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5qh" event={"ID":"f3d120a8-d58e-45b9-be60-04be599332b3","Type":"ContainerDied","Data":"760ee0d53304a79ced9adbeed9c64f4c5f56137c58551b302a474e0402afed15"} Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.578973 4848 generic.go:334] "Generic (PLEG): container finished" podID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerID="8d74ab55ac61323e251f69651a9514773409356a02c83211c6f7d7f8f433d649" exitCode=0 Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.579033 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2dzv" event={"ID":"00f2c84d-1e46-4563-9c78-6657bb601db3","Type":"ContainerDied","Data":"8d74ab55ac61323e251f69651a9514773409356a02c83211c6f7d7f8f433d649"} Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.579196 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2dzv" event={"ID":"00f2c84d-1e46-4563-9c78-6657bb601db3","Type":"ContainerStarted","Data":"1b7d3a6d808b16283114a935c37f75e7abdc9d9af23ec2f3ca291d1f9e377590"} Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.580335 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.581090 4848 generic.go:334] "Generic (PLEG): container finished" podID="e75ab632-da17-4354-951c-476ad2e46835" containerID="c5b4cb4faaea6a1fa269389d85f5f2cbeb9d4d4fd0a8d7aa390102ca8042950c" exitCode=0 Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.581202 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjljj" event={"ID":"e75ab632-da17-4354-951c-476ad2e46835","Type":"ContainerDied","Data":"c5b4cb4faaea6a1fa269389d85f5f2cbeb9d4d4fd0a8d7aa390102ca8042950c"} Dec 04 13:50:53 crc kubenswrapper[4848]: I1204 13:50:53.581229 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjljj" event={"ID":"e75ab632-da17-4354-951c-476ad2e46835","Type":"ContainerStarted","Data":"215f327d5a8796c31cc9dc05de5f1ed6087d8a66b39ffe77470f075e80b3b2d4"} Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.008753 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hrx58"] Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.010265 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.012802 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.022402 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrx58"] Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.200314 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-catalog-content\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.200406 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-utilities\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.200532 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4zvc\" (UniqueName: \"kubernetes.io/projected/8eedf821-9e1a-4513-896f-372df7171ef2-kube-api-access-k4zvc\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.203600 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.261899 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:54 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:54 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:54 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.262000 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.302069 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca4d77ac-6c91-4789-8a01-553fc2563d65-secret-volume\") pod \"ca4d77ac-6c91-4789-8a01-553fc2563d65\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.302240 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca4d77ac-6c91-4789-8a01-553fc2563d65-config-volume\") pod \"ca4d77ac-6c91-4789-8a01-553fc2563d65\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.302407 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6hw5\" (UniqueName: \"kubernetes.io/projected/ca4d77ac-6c91-4789-8a01-553fc2563d65-kube-api-access-k6hw5\") pod \"ca4d77ac-6c91-4789-8a01-553fc2563d65\" (UID: \"ca4d77ac-6c91-4789-8a01-553fc2563d65\") " Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.303467 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca4d77ac-6c91-4789-8a01-553fc2563d65-config-volume" (OuterVolumeSpecName: "config-volume") pod "ca4d77ac-6c91-4789-8a01-553fc2563d65" (UID: "ca4d77ac-6c91-4789-8a01-553fc2563d65"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.304512 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4zvc\" (UniqueName: \"kubernetes.io/projected/8eedf821-9e1a-4513-896f-372df7171ef2-kube-api-access-k4zvc\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.304623 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-catalog-content\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.304727 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-utilities\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.304908 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca4d77ac-6c91-4789-8a01-553fc2563d65-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.305589 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-utilities\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.309368 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4d77ac-6c91-4789-8a01-553fc2563d65-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ca4d77ac-6c91-4789-8a01-553fc2563d65" (UID: "ca4d77ac-6c91-4789-8a01-553fc2563d65"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.309482 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-catalog-content\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.309803 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca4d77ac-6c91-4789-8a01-553fc2563d65-kube-api-access-k6hw5" (OuterVolumeSpecName: "kube-api-access-k6hw5") pod "ca4d77ac-6c91-4789-8a01-553fc2563d65" (UID: "ca4d77ac-6c91-4789-8a01-553fc2563d65"). InnerVolumeSpecName "kube-api-access-k6hw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.348937 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4zvc\" (UniqueName: \"kubernetes.io/projected/8eedf821-9e1a-4513-896f-372df7171ef2-kube-api-access-k4zvc\") pod \"redhat-marketplace-hrx58\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.410011 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6hw5\" (UniqueName: \"kubernetes.io/projected/ca4d77ac-6c91-4789-8a01-553fc2563d65-kube-api-access-k6hw5\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.410364 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ca4d77ac-6c91-4789-8a01-553fc2563d65-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.422493 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.423342 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.434049 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9dprj"] Dec 04 13:50:54 crc kubenswrapper[4848]: E1204 13:50:54.434314 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca4d77ac-6c91-4789-8a01-553fc2563d65" containerName="collect-profiles" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.434328 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca4d77ac-6c91-4789-8a01-553fc2563d65" containerName="collect-profiles" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.434434 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca4d77ac-6c91-4789-8a01-553fc2563d65" containerName="collect-profiles" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.435247 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-q9lhv" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.435336 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.441647 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dprj"] Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.585699 4848 generic.go:334] "Generic (PLEG): container finished" podID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerID="4635ad052766c3821eb228a283ed7226b1460d0f16d8fcc1c1598b89af25e286" exitCode=0 Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.585749 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp8lv" event={"ID":"f98ec77d-021a-43cd-b8e9-12738f4921f1","Type":"ContainerDied","Data":"4635ad052766c3821eb228a283ed7226b1460d0f16d8fcc1c1598b89af25e286"} Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.589588 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" event={"ID":"ca4d77ac-6c91-4789-8a01-553fc2563d65","Type":"ContainerDied","Data":"cf1016bd99276cd90e2ced6a57abd92f1b6b18c50ea7bb619de65e98fb7b9d78"} Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.589614 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf1016bd99276cd90e2ced6a57abd92f1b6b18c50ea7bb619de65e98fb7b9d78" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.589659 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.591919 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" event={"ID":"9b739fda-b42c-49be-99dc-4ff123bb8cb7","Type":"ContainerStarted","Data":"605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594"} Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.592066 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.612133 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tswgx\" (UniqueName: \"kubernetes.io/projected/75519da1-f479-431a-adbb-0713e0049bd5-kube-api-access-tswgx\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.612178 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-catalog-content\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.612293 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-utilities\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.623274 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.654632 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" podStartSLOduration=132.654613525 podStartE2EDuration="2m12.654613525s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:54.644775781 +0000 UTC m=+158.587272309" watchObservedRunningTime="2025-12-04 13:50:54.654613525 +0000 UTC m=+158.597110053" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.712986 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tswgx\" (UniqueName: \"kubernetes.io/projected/75519da1-f479-431a-adbb-0713e0049bd5-kube-api-access-tswgx\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.713030 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-catalog-content\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.713442 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-catalog-content\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.713577 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-utilities\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.714142 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-utilities\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.757027 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tswgx\" (UniqueName: \"kubernetes.io/projected/75519da1-f479-431a-adbb-0713e0049bd5-kube-api-access-tswgx\") pod \"redhat-marketplace-9dprj\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.765270 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.917819 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrx58"] Dec 04 13:50:54 crc kubenswrapper[4848]: I1204 13:50:54.979827 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dprj"] Dec 04 13:50:54 crc kubenswrapper[4848]: W1204 13:50:54.999472 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75519da1_f479_431a_adbb_0713e0049bd5.slice/crio-e0cbef3b84553ada09448102c239a872adbbf78dadd67bc5ee4d72fb7357cbbf WatchSource:0}: Error finding container e0cbef3b84553ada09448102c239a872adbbf78dadd67bc5ee4d72fb7357cbbf: Status 404 returned error can't find the container with id e0cbef3b84553ada09448102c239a872adbbf78dadd67bc5ee4d72fb7357cbbf Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.224175 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x6h7q"] Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.230906 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.231846 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x6h7q"] Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.233548 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.260633 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:55 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:55 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:55 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.260920 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.328276 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-utilities\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.328341 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-catalog-content\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.328421 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89jwl\" (UniqueName: \"kubernetes.io/projected/15f8738d-4240-42d0-a819-cccff5aaa821-kube-api-access-89jwl\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.349437 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.350391 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.353814 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.361534 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.361759 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.389122 4848 patch_prober.go:28] interesting pod/downloads-7954f5f757-47fkv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.389161 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-47fkv" podUID="9ba41edc-1966-4bca-ae21-b80cdd48ea80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.389505 4848 patch_prober.go:28] interesting pod/downloads-7954f5f757-47fkv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.389523 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-47fkv" podUID="9ba41edc-1966-4bca-ae21-b80cdd48ea80" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.390583 4848 patch_prober.go:28] interesting pod/console-f9d7485db-rrtwl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.390605 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rrtwl" podUID="d175633c-17b1-48b5-a1ed-430f08118b6c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.392019 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.392073 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.430815 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-catalog-content\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.430888 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89jwl\" (UniqueName: \"kubernetes.io/projected/15f8738d-4240-42d0-a819-cccff5aaa821-kube-api-access-89jwl\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.430927 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.430973 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.430991 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-utilities\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.431438 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-utilities\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.431668 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-catalog-content\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.449463 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89jwl\" (UniqueName: \"kubernetes.io/projected/15f8738d-4240-42d0-a819-cccff5aaa821-kube-api-access-89jwl\") pod \"redhat-operators-x6h7q\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.532602 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.532690 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.533353 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.547541 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.606401 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r8vn4"] Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.610154 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.614847 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r8vn4"] Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.616061 4848 generic.go:334] "Generic (PLEG): container finished" podID="75519da1-f479-431a-adbb-0713e0049bd5" containerID="57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157" exitCode=0 Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.616194 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dprj" event={"ID":"75519da1-f479-431a-adbb-0713e0049bd5","Type":"ContainerDied","Data":"57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157"} Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.616227 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dprj" event={"ID":"75519da1-f479-431a-adbb-0713e0049bd5","Type":"ContainerStarted","Data":"e0cbef3b84553ada09448102c239a872adbbf78dadd67bc5ee4d72fb7357cbbf"} Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.623125 4848 generic.go:334] "Generic (PLEG): container finished" podID="8eedf821-9e1a-4513-896f-372df7171ef2" containerID="33c698f047dbcb19bd2cec40862123a70ff599ccb99d9490a551bedc163f26b4" exitCode=0 Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.624106 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrx58" event={"ID":"8eedf821-9e1a-4513-896f-372df7171ef2","Type":"ContainerDied","Data":"33c698f047dbcb19bd2cec40862123a70ff599ccb99d9490a551bedc163f26b4"} Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.624129 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrx58" event={"ID":"8eedf821-9e1a-4513-896f-372df7171ef2","Type":"ContainerStarted","Data":"9914984ebb7744afd30a678f6134757d9147b5adc576b03e36a767090d4975f9"} Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.735751 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-catalog-content\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.735794 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnzzw\" (UniqueName: \"kubernetes.io/projected/5949452c-a2a6-488a-8730-2edb4cfc5747-kube-api-access-xnzzw\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.735815 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-utilities\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.824442 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x6h7q"] Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.837322 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-catalog-content\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.837567 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnzzw\" (UniqueName: \"kubernetes.io/projected/5949452c-a2a6-488a-8730-2edb4cfc5747-kube-api-access-xnzzw\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.837588 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-utilities\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.837918 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-utilities\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.838596 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-catalog-content\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.858003 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnzzw\" (UniqueName: \"kubernetes.io/projected/5949452c-a2a6-488a-8730-2edb4cfc5747-kube-api-access-xnzzw\") pod \"redhat-operators-r8vn4\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:55 crc kubenswrapper[4848]: I1204 13:50:55.966673 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.255251 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r8vn4"] Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.258085 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.260406 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.263088 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:56 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:56 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:56 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.263130 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.307251 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.373147 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.373815 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.379573 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.379824 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.380072 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.446237 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.446303 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.547244 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.547309 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.547381 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.577072 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.613297 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.630088 4848 generic.go:334] "Generic (PLEG): container finished" podID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerID="b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac" exitCode=0 Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.630400 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r8vn4" event={"ID":"5949452c-a2a6-488a-8730-2edb4cfc5747","Type":"ContainerDied","Data":"b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac"} Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.631262 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r8vn4" event={"ID":"5949452c-a2a6-488a-8730-2edb4cfc5747","Type":"ContainerStarted","Data":"22f29693a6454a29ae4effcaf52be2d69834f612e9e3660c4d2add5df9709ee4"} Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.633618 4848 generic.go:334] "Generic (PLEG): container finished" podID="15f8738d-4240-42d0-a819-cccff5aaa821" containerID="fa918e58186f5974a99e08cd4471f2311711c21a8e366c82b241acb4d6b48537" exitCode=0 Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.633638 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6h7q" event={"ID":"15f8738d-4240-42d0-a819-cccff5aaa821","Type":"ContainerDied","Data":"fa918e58186f5974a99e08cd4471f2311711c21a8e366c82b241acb4d6b48537"} Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.633654 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6h7q" event={"ID":"15f8738d-4240-42d0-a819-cccff5aaa821","Type":"ContainerStarted","Data":"dd10f453727a3f8bab92f925578ca42bb79d22e440d3e3fdd271deda31cf114d"} Dec 04 13:50:56 crc kubenswrapper[4848]: I1204 13:50:56.854320 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:50:57 crc kubenswrapper[4848]: I1204 13:50:57.182695 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 13:50:57 crc kubenswrapper[4848]: W1204 13:50:57.221002 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9b75b38d_dbb9_4942_a1c2_7a51059eaf0d.slice/crio-adee3e457347ccba350f7f4436d6f85533f7626126eba8fef4f81b5b94739e20 WatchSource:0}: Error finding container adee3e457347ccba350f7f4436d6f85533f7626126eba8fef4f81b5b94739e20: Status 404 returned error can't find the container with id adee3e457347ccba350f7f4436d6f85533f7626126eba8fef4f81b5b94739e20 Dec 04 13:50:57 crc kubenswrapper[4848]: I1204 13:50:57.261168 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:57 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:57 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:57 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:57 crc kubenswrapper[4848]: I1204 13:50:57.261216 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:57 crc kubenswrapper[4848]: I1204 13:50:57.677601 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d","Type":"ContainerStarted","Data":"adee3e457347ccba350f7f4436d6f85533f7626126eba8fef4f81b5b94739e20"} Dec 04 13:50:57 crc kubenswrapper[4848]: I1204 13:50:57.690448 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"72d2fc29-f973-4778-aed2-9b000ea7e4bc","Type":"ContainerStarted","Data":"5076f8c5ceb8b1d8e7fc395efdba0b11aa8ed6c097fa9fe24000b2228bb255c4"} Dec 04 13:50:57 crc kubenswrapper[4848]: I1204 13:50:57.690491 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"72d2fc29-f973-4778-aed2-9b000ea7e4bc","Type":"ContainerStarted","Data":"a7bf0a80dc9954f81d49f0c59c0f66733275497e83c939392a30d1e725b27137"} Dec 04 13:50:57 crc kubenswrapper[4848]: I1204 13:50:57.710319 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.710263268 podStartE2EDuration="2.710263268s" podCreationTimestamp="2025-12-04 13:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:50:57.70629531 +0000 UTC m=+161.648791838" watchObservedRunningTime="2025-12-04 13:50:57.710263268 +0000 UTC m=+161.652759796" Dec 04 13:50:58 crc kubenswrapper[4848]: I1204 13:50:58.262766 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:58 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:58 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:58 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:58 crc kubenswrapper[4848]: I1204 13:50:58.262816 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:58 crc kubenswrapper[4848]: I1204 13:50:58.705655 4848 generic.go:334] "Generic (PLEG): container finished" podID="72d2fc29-f973-4778-aed2-9b000ea7e4bc" containerID="5076f8c5ceb8b1d8e7fc395efdba0b11aa8ed6c097fa9fe24000b2228bb255c4" exitCode=0 Dec 04 13:50:58 crc kubenswrapper[4848]: I1204 13:50:58.705720 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"72d2fc29-f973-4778-aed2-9b000ea7e4bc","Type":"ContainerDied","Data":"5076f8c5ceb8b1d8e7fc395efdba0b11aa8ed6c097fa9fe24000b2228bb255c4"} Dec 04 13:50:58 crc kubenswrapper[4848]: I1204 13:50:58.712607 4848 generic.go:334] "Generic (PLEG): container finished" podID="9b75b38d-dbb9-4942-a1c2-7a51059eaf0d" containerID="234364dac5585a40ca9458f75c955a94f1d36081d996d67f44d69b2c65c3a2c4" exitCode=0 Dec 04 13:50:58 crc kubenswrapper[4848]: I1204 13:50:58.712643 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d","Type":"ContainerDied","Data":"234364dac5585a40ca9458f75c955a94f1d36081d996d67f44d69b2c65c3a2c4"} Dec 04 13:50:58 crc kubenswrapper[4848]: I1204 13:50:58.823495 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 13:50:59 crc kubenswrapper[4848]: I1204 13:50:59.260784 4848 patch_prober.go:28] interesting pod/router-default-5444994796-7crwb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 13:50:59 crc kubenswrapper[4848]: [-]has-synced failed: reason withheld Dec 04 13:50:59 crc kubenswrapper[4848]: [+]process-running ok Dec 04 13:50:59 crc kubenswrapper[4848]: healthz check failed Dec 04 13:50:59 crc kubenswrapper[4848]: I1204 13:50:59.261303 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7crwb" podUID="8ccad0f8-a334-4847-8c78-56a2d3b81507" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:50:59 crc kubenswrapper[4848]: I1204 13:50:59.684079 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-nftbz" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.040626 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.153355 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kubelet-dir\") pod \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\" (UID: \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\") " Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.153417 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kube-api-access\") pod \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\" (UID: \"72d2fc29-f973-4778-aed2-9b000ea7e4bc\") " Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.153539 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "72d2fc29-f973-4778-aed2-9b000ea7e4bc" (UID: "72d2fc29-f973-4778-aed2-9b000ea7e4bc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.154816 4848 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.159466 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "72d2fc29-f973-4778-aed2-9b000ea7e4bc" (UID: "72d2fc29-f973-4778-aed2-9b000ea7e4bc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.214293 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.255401 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kubelet-dir\") pod \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\" (UID: \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\") " Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.255476 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kube-api-access\") pod \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\" (UID: \"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d\") " Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.255514 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9b75b38d-dbb9-4942-a1c2-7a51059eaf0d" (UID: "9b75b38d-dbb9-4942-a1c2-7a51059eaf0d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.255763 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/72d2fc29-f973-4778-aed2-9b000ea7e4bc-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.255781 4848 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.258341 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9b75b38d-dbb9-4942-a1c2-7a51059eaf0d" (UID: "9b75b38d-dbb9-4942-a1c2-7a51059eaf0d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.265217 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.267442 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7crwb" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.358587 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b75b38d-dbb9-4942-a1c2-7a51059eaf0d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.750126 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b75b38d-dbb9-4942-a1c2-7a51059eaf0d","Type":"ContainerDied","Data":"adee3e457347ccba350f7f4436d6f85533f7626126eba8fef4f81b5b94739e20"} Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.750165 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adee3e457347ccba350f7f4436d6f85533f7626126eba8fef4f81b5b94739e20" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.750224 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.753810 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.753873 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"72d2fc29-f973-4778-aed2-9b000ea7e4bc","Type":"ContainerDied","Data":"a7bf0a80dc9954f81d49f0c59c0f66733275497e83c939392a30d1e725b27137"} Dec 04 13:51:00 crc kubenswrapper[4848]: I1204 13:51:00.753909 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7bf0a80dc9954f81d49f0c59c0f66733275497e83c939392a30d1e725b27137" Dec 04 13:51:04 crc kubenswrapper[4848]: I1204 13:51:04.886196 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:51:04 crc kubenswrapper[4848]: I1204 13:51:04.955263 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5577aeec-1fac-4adc-a566-b0110dd2477b-metrics-certs\") pod \"network-metrics-daemon-v4k6x\" (UID: \"5577aeec-1fac-4adc-a566-b0110dd2477b\") " pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:51:05 crc kubenswrapper[4848]: I1204 13:51:05.011839 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-v4k6x" Dec 04 13:51:05 crc kubenswrapper[4848]: I1204 13:51:05.393052 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:51:05 crc kubenswrapper[4848]: I1204 13:51:05.400333 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:51:05 crc kubenswrapper[4848]: I1204 13:51:05.403847 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-47fkv" Dec 04 13:51:13 crc kubenswrapper[4848]: I1204 13:51:13.455908 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:51:14 crc kubenswrapper[4848]: I1204 13:51:14.314660 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:51:14 crc kubenswrapper[4848]: I1204 13:51:14.314727 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:51:26 crc kubenswrapper[4848]: I1204 13:51:26.313262 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rqjh6" Dec 04 13:51:26 crc kubenswrapper[4848]: I1204 13:51:26.756412 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.749860 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 13:51:30 crc kubenswrapper[4848]: E1204 13:51:30.750639 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b75b38d-dbb9-4942-a1c2-7a51059eaf0d" containerName="pruner" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.750654 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b75b38d-dbb9-4942-a1c2-7a51059eaf0d" containerName="pruner" Dec 04 13:51:30 crc kubenswrapper[4848]: E1204 13:51:30.750670 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d2fc29-f973-4778-aed2-9b000ea7e4bc" containerName="pruner" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.750678 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d2fc29-f973-4778-aed2-9b000ea7e4bc" containerName="pruner" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.750815 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b75b38d-dbb9-4942-a1c2-7a51059eaf0d" containerName="pruner" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.750835 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d2fc29-f973-4778-aed2-9b000ea7e4bc" containerName="pruner" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.751284 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.760399 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.761343 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.765093 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.932743 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60783e1c-df42-4e49-90d9-1aa0a1fef050-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"60783e1c-df42-4e49-90d9-1aa0a1fef050\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:30 crc kubenswrapper[4848]: I1204 13:51:30.932787 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60783e1c-df42-4e49-90d9-1aa0a1fef050-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"60783e1c-df42-4e49-90d9-1aa0a1fef050\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:31 crc kubenswrapper[4848]: I1204 13:51:31.033900 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60783e1c-df42-4e49-90d9-1aa0a1fef050-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"60783e1c-df42-4e49-90d9-1aa0a1fef050\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:31 crc kubenswrapper[4848]: I1204 13:51:31.033945 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60783e1c-df42-4e49-90d9-1aa0a1fef050-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"60783e1c-df42-4e49-90d9-1aa0a1fef050\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:31 crc kubenswrapper[4848]: I1204 13:51:31.034059 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60783e1c-df42-4e49-90d9-1aa0a1fef050-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"60783e1c-df42-4e49-90d9-1aa0a1fef050\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:31 crc kubenswrapper[4848]: I1204 13:51:31.052069 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60783e1c-df42-4e49-90d9-1aa0a1fef050-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"60783e1c-df42-4e49-90d9-1aa0a1fef050\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:31 crc kubenswrapper[4848]: I1204 13:51:31.084017 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:34 crc kubenswrapper[4848]: E1204 13:51:34.329019 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 13:51:34 crc kubenswrapper[4848]: E1204 13:51:34.329512 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mkqg2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kp8lv_openshift-marketplace(f98ec77d-021a-43cd-b8e9-12738f4921f1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 13:51:34 crc kubenswrapper[4848]: E1204 13:51:34.330918 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kp8lv" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" Dec 04 13:51:35 crc kubenswrapper[4848]: E1204 13:51:35.485832 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kp8lv" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" Dec 04 13:51:35 crc kubenswrapper[4848]: E1204 13:51:35.733224 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 13:51:35 crc kubenswrapper[4848]: E1204 13:51:35.733536 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tswgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-9dprj_openshift-marketplace(75519da1-f479-431a-adbb-0713e0049bd5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 13:51:35 crc kubenswrapper[4848]: E1204 13:51:35.734862 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-9dprj" podUID="75519da1-f479-431a-adbb-0713e0049bd5" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.142657 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.144024 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.150096 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.307211 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kube-api-access\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.307375 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-var-lock\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.307656 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.408987 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-var-lock\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.409038 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.409075 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kube-api-access\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.409464 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-var-lock\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.409536 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.432408 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kube-api-access\") pod \"installer-9-crc\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:36 crc kubenswrapper[4848]: I1204 13:51:36.468185 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:51:37 crc kubenswrapper[4848]: E1204 13:51:37.446163 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-9dprj" podUID="75519da1-f479-431a-adbb-0713e0049bd5" Dec 04 13:51:37 crc kubenswrapper[4848]: E1204 13:51:37.566869 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 13:51:37 crc kubenswrapper[4848]: E1204 13:51:37.567337 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wfblq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-c2dzv_openshift-marketplace(00f2c84d-1e46-4563-9c78-6657bb601db3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 13:51:37 crc kubenswrapper[4848]: E1204 13:51:37.568645 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-c2dzv" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" Dec 04 13:51:44 crc kubenswrapper[4848]: E1204 13:51:44.210096 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-c2dzv" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" Dec 04 13:51:44 crc kubenswrapper[4848]: I1204 13:51:44.314094 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:51:44 crc kubenswrapper[4848]: I1204 13:51:44.314344 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:51:44 crc kubenswrapper[4848]: I1204 13:51:44.314386 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:51:44 crc kubenswrapper[4848]: I1204 13:51:44.314781 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:51:44 crc kubenswrapper[4848]: I1204 13:51:44.314877 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1" gracePeriod=600 Dec 04 13:51:44 crc kubenswrapper[4848]: I1204 13:51:44.432522 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-v4k6x"] Dec 04 13:51:44 crc kubenswrapper[4848]: I1204 13:51:44.616161 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 13:51:44 crc kubenswrapper[4848]: I1204 13:51:44.625321 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 13:51:44 crc kubenswrapper[4848]: W1204 13:51:44.658301 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0a33ea42_e76c_4853_b7ad_68f96192d8f4.slice/crio-e97e006ea342c1cb897f9b4d53e4ed71b80a99b7f76ea5140d5a7288b93aca90 WatchSource:0}: Error finding container e97e006ea342c1cb897f9b4d53e4ed71b80a99b7f76ea5140d5a7288b93aca90: Status 404 returned error can't find the container with id e97e006ea342c1cb897f9b4d53e4ed71b80a99b7f76ea5140d5a7288b93aca90 Dec 04 13:51:45 crc kubenswrapper[4848]: I1204 13:51:45.051080 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" event={"ID":"5577aeec-1fac-4adc-a566-b0110dd2477b","Type":"ContainerStarted","Data":"90ce20eaf14a83abf72a672de6ef021bce750be5706eb9bfb33a94fdc5f9e4cf"} Dec 04 13:51:45 crc kubenswrapper[4848]: I1204 13:51:45.051902 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"60783e1c-df42-4e49-90d9-1aa0a1fef050","Type":"ContainerStarted","Data":"bd34edf9efd39f915ef4ddd15526f9da2783cee5da7cd7110a2786e484ec3825"} Dec 04 13:51:45 crc kubenswrapper[4848]: I1204 13:51:45.052920 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0a33ea42-e76c-4853-b7ad-68f96192d8f4","Type":"ContainerStarted","Data":"e97e006ea342c1cb897f9b4d53e4ed71b80a99b7f76ea5140d5a7288b93aca90"} Dec 04 13:51:47 crc kubenswrapper[4848]: I1204 13:51:47.067897 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0a33ea42-e76c-4853-b7ad-68f96192d8f4","Type":"ContainerStarted","Data":"2382242c375993091efabeae4815a7c23bc6d65e947f25d730662d6dba6ba24c"} Dec 04 13:51:47 crc kubenswrapper[4848]: I1204 13:51:47.071365 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1" exitCode=0 Dec 04 13:51:47 crc kubenswrapper[4848]: I1204 13:51:47.071405 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1"} Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.853364 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.853966 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-89jwl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x6h7q_openshift-marketplace(15f8738d-4240-42d0-a819-cccff5aaa821): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.854509 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.854702 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xnzzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-r8vn4_openshift-marketplace(5949452c-a2a6-488a-8730-2edb4cfc5747): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.855555 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-x6h7q" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.857639 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-r8vn4" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.859984 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.860145 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lgbzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-sjljj_openshift-marketplace(e75ab632-da17-4354-951c-476ad2e46835): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.861985 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-sjljj" podUID="e75ab632-da17-4354-951c-476ad2e46835" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.881894 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.882029 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fgk26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nr5qh_openshift-marketplace(f3d120a8-d58e-45b9-be60-04be599332b3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.883439 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nr5qh" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.929035 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.929175 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k4zvc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hrx58_openshift-marketplace(8eedf821-9e1a-4513-896f-372df7171ef2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 13:51:47 crc kubenswrapper[4848]: E1204 13:51:47.930340 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hrx58" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" Dec 04 13:51:48 crc kubenswrapper[4848]: I1204 13:51:48.079809 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"60783e1c-df42-4e49-90d9-1aa0a1fef050","Type":"ContainerStarted","Data":"a949e30a3e5c5d1622b2a596514dfe3996cf1f9a1fc46a09b0cbe2cd281f75ad"} Dec 04 13:51:48 crc kubenswrapper[4848]: I1204 13:51:48.082530 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"72d765596d3c5a2d39e725591250e92b81042c983d061e6907f508172e90aa2a"} Dec 04 13:51:48 crc kubenswrapper[4848]: I1204 13:51:48.085687 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" event={"ID":"5577aeec-1fac-4adc-a566-b0110dd2477b","Type":"ContainerStarted","Data":"13868f16e101c24e5d2c39734648bf3706edd6ce393a2ec7d3b41e9db204b24a"} Dec 04 13:51:48 crc kubenswrapper[4848]: I1204 13:51:48.085721 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-v4k6x" event={"ID":"5577aeec-1fac-4adc-a566-b0110dd2477b","Type":"ContainerStarted","Data":"1921b824d7d661521d3bb31f68cdd835e3e21d966e4620cecfa011a4a2f1328a"} Dec 04 13:51:48 crc kubenswrapper[4848]: E1204 13:51:48.087443 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-r8vn4" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" Dec 04 13:51:48 crc kubenswrapper[4848]: E1204 13:51:48.087693 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hrx58" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" Dec 04 13:51:48 crc kubenswrapper[4848]: E1204 13:51:48.087777 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x6h7q" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" Dec 04 13:51:48 crc kubenswrapper[4848]: E1204 13:51:48.087808 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nr5qh" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" Dec 04 13:51:48 crc kubenswrapper[4848]: E1204 13:51:48.087968 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-sjljj" podUID="e75ab632-da17-4354-951c-476ad2e46835" Dec 04 13:51:48 crc kubenswrapper[4848]: I1204 13:51:48.116289 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=18.11625769 podStartE2EDuration="18.11625769s" podCreationTimestamp="2025-12-04 13:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:51:48.09538527 +0000 UTC m=+212.037881818" watchObservedRunningTime="2025-12-04 13:51:48.11625769 +0000 UTC m=+212.058754258" Dec 04 13:51:48 crc kubenswrapper[4848]: E1204 13:51:48.198783 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod60783e1c_df42_4e49_90d9_1aa0a1fef050.slice/crio-a949e30a3e5c5d1622b2a596514dfe3996cf1f9a1fc46a09b0cbe2cd281f75ad.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-pod60783e1c_df42_4e49_90d9_1aa0a1fef050.slice/crio-conmon-a949e30a3e5c5d1622b2a596514dfe3996cf1f9a1fc46a09b0cbe2cd281f75ad.scope\": RecentStats: unable to find data in memory cache]" Dec 04 13:51:48 crc kubenswrapper[4848]: I1204 13:51:48.234644 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=12.234625493 podStartE2EDuration="12.234625493s" podCreationTimestamp="2025-12-04 13:51:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:51:48.233486174 +0000 UTC m=+212.175982702" watchObservedRunningTime="2025-12-04 13:51:48.234625493 +0000 UTC m=+212.177122021" Dec 04 13:51:48 crc kubenswrapper[4848]: I1204 13:51:48.263753 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-v4k6x" podStartSLOduration=186.263739608 podStartE2EDuration="3m6.263739608s" podCreationTimestamp="2025-12-04 13:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:51:48.261293278 +0000 UTC m=+212.203789806" watchObservedRunningTime="2025-12-04 13:51:48.263739608 +0000 UTC m=+212.206236136" Dec 04 13:51:49 crc kubenswrapper[4848]: I1204 13:51:49.092553 4848 generic.go:334] "Generic (PLEG): container finished" podID="60783e1c-df42-4e49-90d9-1aa0a1fef050" containerID="a949e30a3e5c5d1622b2a596514dfe3996cf1f9a1fc46a09b0cbe2cd281f75ad" exitCode=0 Dec 04 13:51:49 crc kubenswrapper[4848]: I1204 13:51:49.092635 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"60783e1c-df42-4e49-90d9-1aa0a1fef050","Type":"ContainerDied","Data":"a949e30a3e5c5d1622b2a596514dfe3996cf1f9a1fc46a09b0cbe2cd281f75ad"} Dec 04 13:51:50 crc kubenswrapper[4848]: I1204 13:51:50.329799 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:50 crc kubenswrapper[4848]: I1204 13:51:50.396366 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60783e1c-df42-4e49-90d9-1aa0a1fef050-kube-api-access\") pod \"60783e1c-df42-4e49-90d9-1aa0a1fef050\" (UID: \"60783e1c-df42-4e49-90d9-1aa0a1fef050\") " Dec 04 13:51:50 crc kubenswrapper[4848]: I1204 13:51:50.397168 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60783e1c-df42-4e49-90d9-1aa0a1fef050-kubelet-dir\") pod \"60783e1c-df42-4e49-90d9-1aa0a1fef050\" (UID: \"60783e1c-df42-4e49-90d9-1aa0a1fef050\") " Dec 04 13:51:50 crc kubenswrapper[4848]: I1204 13:51:50.397371 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60783e1c-df42-4e49-90d9-1aa0a1fef050-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "60783e1c-df42-4e49-90d9-1aa0a1fef050" (UID: "60783e1c-df42-4e49-90d9-1aa0a1fef050"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:51:50 crc kubenswrapper[4848]: I1204 13:51:50.403199 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60783e1c-df42-4e49-90d9-1aa0a1fef050-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "60783e1c-df42-4e49-90d9-1aa0a1fef050" (UID: "60783e1c-df42-4e49-90d9-1aa0a1fef050"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:51:50 crc kubenswrapper[4848]: I1204 13:51:50.498346 4848 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/60783e1c-df42-4e49-90d9-1aa0a1fef050-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:50 crc kubenswrapper[4848]: I1204 13:51:50.498382 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60783e1c-df42-4e49-90d9-1aa0a1fef050-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:51 crc kubenswrapper[4848]: I1204 13:51:51.109173 4848 generic.go:334] "Generic (PLEG): container finished" podID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerID="47893295eba47e208843c9bd972e6af31a4fa0f3eb7edfa12f35fd970cd36da8" exitCode=0 Dec 04 13:51:51 crc kubenswrapper[4848]: I1204 13:51:51.109267 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp8lv" event={"ID":"f98ec77d-021a-43cd-b8e9-12738f4921f1","Type":"ContainerDied","Data":"47893295eba47e208843c9bd972e6af31a4fa0f3eb7edfa12f35fd970cd36da8"} Dec 04 13:51:51 crc kubenswrapper[4848]: I1204 13:51:51.114347 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"60783e1c-df42-4e49-90d9-1aa0a1fef050","Type":"ContainerDied","Data":"bd34edf9efd39f915ef4ddd15526f9da2783cee5da7cd7110a2786e484ec3825"} Dec 04 13:51:51 crc kubenswrapper[4848]: I1204 13:51:51.117100 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd34edf9efd39f915ef4ddd15526f9da2783cee5da7cd7110a2786e484ec3825" Dec 04 13:51:51 crc kubenswrapper[4848]: I1204 13:51:51.115150 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 13:51:51 crc kubenswrapper[4848]: I1204 13:51:51.121081 4848 generic.go:334] "Generic (PLEG): container finished" podID="75519da1-f479-431a-adbb-0713e0049bd5" containerID="b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463" exitCode=0 Dec 04 13:51:51 crc kubenswrapper[4848]: I1204 13:51:51.121102 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dprj" event={"ID":"75519da1-f479-431a-adbb-0713e0049bd5","Type":"ContainerDied","Data":"b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463"} Dec 04 13:51:52 crc kubenswrapper[4848]: I1204 13:51:52.129480 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp8lv" event={"ID":"f98ec77d-021a-43cd-b8e9-12738f4921f1","Type":"ContainerStarted","Data":"28edbc3e80b44ef66937f62510bc8f12542c7f2a7f52ac597af4209b0b8a520e"} Dec 04 13:51:52 crc kubenswrapper[4848]: I1204 13:51:52.132549 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dprj" event={"ID":"75519da1-f479-431a-adbb-0713e0049bd5","Type":"ContainerStarted","Data":"3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8"} Dec 04 13:51:52 crc kubenswrapper[4848]: I1204 13:51:52.150584 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kp8lv" podStartSLOduration=3.003090267 podStartE2EDuration="1m0.150561793s" podCreationTimestamp="2025-12-04 13:50:52 +0000 UTC" firstStartedPulling="2025-12-04 13:50:54.587434911 +0000 UTC m=+158.529931439" lastFinishedPulling="2025-12-04 13:51:51.734906427 +0000 UTC m=+215.677402965" observedRunningTime="2025-12-04 13:51:52.148817309 +0000 UTC m=+216.091313867" watchObservedRunningTime="2025-12-04 13:51:52.150561793 +0000 UTC m=+216.093058341" Dec 04 13:51:52 crc kubenswrapper[4848]: I1204 13:51:52.170516 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9dprj" podStartSLOduration=2.281705305 podStartE2EDuration="58.17049692s" podCreationTimestamp="2025-12-04 13:50:54 +0000 UTC" firstStartedPulling="2025-12-04 13:50:55.618380732 +0000 UTC m=+159.560877260" lastFinishedPulling="2025-12-04 13:51:51.507172347 +0000 UTC m=+215.449668875" observedRunningTime="2025-12-04 13:51:52.169059964 +0000 UTC m=+216.111556512" watchObservedRunningTime="2025-12-04 13:51:52.17049692 +0000 UTC m=+216.112993448" Dec 04 13:51:52 crc kubenswrapper[4848]: I1204 13:51:52.972279 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:51:52 crc kubenswrapper[4848]: I1204 13:51:52.972670 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:51:54 crc kubenswrapper[4848]: I1204 13:51:54.056055 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kp8lv" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="registry-server" probeResult="failure" output=< Dec 04 13:51:54 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 13:51:54 crc kubenswrapper[4848]: > Dec 04 13:51:54 crc kubenswrapper[4848]: I1204 13:51:54.766463 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:51:54 crc kubenswrapper[4848]: I1204 13:51:54.767348 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:51:54 crc kubenswrapper[4848]: I1204 13:51:54.834422 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:51:56 crc kubenswrapper[4848]: I1204 13:51:56.215096 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.181881 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2dzv" event={"ID":"00f2c84d-1e46-4563-9c78-6657bb601db3","Type":"ContainerStarted","Data":"2c1fb7a81aca3ff4e2b951c018a99f7ce6e0056646fc8aae4c3dce2b64bfc35d"} Dec 04 13:51:58 crc kubenswrapper[4848]: E1204 13:51:58.306638 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00f2c84d_1e46_4563_9c78_6657bb601db3.slice/crio-2c1fb7a81aca3ff4e2b951c018a99f7ce6e0056646fc8aae4c3dce2b64bfc35d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00f2c84d_1e46_4563_9c78_6657bb601db3.slice/crio-conmon-2c1fb7a81aca3ff4e2b951c018a99f7ce6e0056646fc8aae4c3dce2b64bfc35d.scope\": RecentStats: unable to find data in memory cache]" Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.326722 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dprj"] Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.326928 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9dprj" podUID="75519da1-f479-431a-adbb-0713e0049bd5" containerName="registry-server" containerID="cri-o://3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8" gracePeriod=2 Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.627922 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.719031 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-utilities\") pod \"75519da1-f479-431a-adbb-0713e0049bd5\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.719095 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tswgx\" (UniqueName: \"kubernetes.io/projected/75519da1-f479-431a-adbb-0713e0049bd5-kube-api-access-tswgx\") pod \"75519da1-f479-431a-adbb-0713e0049bd5\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.719212 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-catalog-content\") pod \"75519da1-f479-431a-adbb-0713e0049bd5\" (UID: \"75519da1-f479-431a-adbb-0713e0049bd5\") " Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.720070 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-utilities" (OuterVolumeSpecName: "utilities") pod "75519da1-f479-431a-adbb-0713e0049bd5" (UID: "75519da1-f479-431a-adbb-0713e0049bd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.726102 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75519da1-f479-431a-adbb-0713e0049bd5-kube-api-access-tswgx" (OuterVolumeSpecName: "kube-api-access-tswgx") pod "75519da1-f479-431a-adbb-0713e0049bd5" (UID: "75519da1-f479-431a-adbb-0713e0049bd5"). InnerVolumeSpecName "kube-api-access-tswgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.759034 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75519da1-f479-431a-adbb-0713e0049bd5" (UID: "75519da1-f479-431a-adbb-0713e0049bd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.820616 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.820650 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tswgx\" (UniqueName: \"kubernetes.io/projected/75519da1-f479-431a-adbb-0713e0049bd5-kube-api-access-tswgx\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:58 crc kubenswrapper[4848]: I1204 13:51:58.820664 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75519da1-f479-431a-adbb-0713e0049bd5-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.187851 4848 generic.go:334] "Generic (PLEG): container finished" podID="75519da1-f479-431a-adbb-0713e0049bd5" containerID="3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8" exitCode=0 Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.187926 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9dprj" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.187928 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dprj" event={"ID":"75519da1-f479-431a-adbb-0713e0049bd5","Type":"ContainerDied","Data":"3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8"} Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.189578 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9dprj" event={"ID":"75519da1-f479-431a-adbb-0713e0049bd5","Type":"ContainerDied","Data":"e0cbef3b84553ada09448102c239a872adbbf78dadd67bc5ee4d72fb7357cbbf"} Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.189604 4848 scope.go:117] "RemoveContainer" containerID="3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.191877 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrx58" event={"ID":"8eedf821-9e1a-4513-896f-372df7171ef2","Type":"ContainerStarted","Data":"3aef13f92912d30feed58b4eca566026cbb430ea49de87ad4bf8fc98f384a926"} Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.193849 4848 generic.go:334] "Generic (PLEG): container finished" podID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerID="2c1fb7a81aca3ff4e2b951c018a99f7ce6e0056646fc8aae4c3dce2b64bfc35d" exitCode=0 Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.193902 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2dzv" event={"ID":"00f2c84d-1e46-4563-9c78-6657bb601db3","Type":"ContainerDied","Data":"2c1fb7a81aca3ff4e2b951c018a99f7ce6e0056646fc8aae4c3dce2b64bfc35d"} Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.204021 4848 scope.go:117] "RemoveContainer" containerID="b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.232261 4848 scope.go:117] "RemoveContainer" containerID="57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.247698 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dprj"] Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.253765 4848 scope.go:117] "RemoveContainer" containerID="3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.254102 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9dprj"] Dec 04 13:51:59 crc kubenswrapper[4848]: E1204 13:51:59.254209 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8\": container with ID starting with 3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8 not found: ID does not exist" containerID="3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.254297 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8"} err="failed to get container status \"3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8\": rpc error: code = NotFound desc = could not find container \"3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8\": container with ID starting with 3115ccb8585afd8861bd00124b6ef237df8fd012265ca7b97b8cc2178ab0eef8 not found: ID does not exist" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.254400 4848 scope.go:117] "RemoveContainer" containerID="b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463" Dec 04 13:51:59 crc kubenswrapper[4848]: E1204 13:51:59.254835 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463\": container with ID starting with b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463 not found: ID does not exist" containerID="b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.254900 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463"} err="failed to get container status \"b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463\": rpc error: code = NotFound desc = could not find container \"b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463\": container with ID starting with b6c08483ece8c55ad24845ff5004688b5038649c971369f2dd1904b568a3f463 not found: ID does not exist" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.254941 4848 scope.go:117] "RemoveContainer" containerID="57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157" Dec 04 13:51:59 crc kubenswrapper[4848]: E1204 13:51:59.255388 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157\": container with ID starting with 57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157 not found: ID does not exist" containerID="57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157" Dec 04 13:51:59 crc kubenswrapper[4848]: I1204 13:51:59.255540 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157"} err="failed to get container status \"57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157\": rpc error: code = NotFound desc = could not find container \"57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157\": container with ID starting with 57fafd3415d38250e63fa59ac3e76cab2e07e3306e29f8da3ce7e08e191c3157 not found: ID does not exist" Dec 04 13:52:00 crc kubenswrapper[4848]: I1204 13:52:00.207863 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2dzv" event={"ID":"00f2c84d-1e46-4563-9c78-6657bb601db3","Type":"ContainerStarted","Data":"a2543378f8bc40ef105c55cd3433908fbf6f08574737dd98bf4ea4c833c9a086"} Dec 04 13:52:00 crc kubenswrapper[4848]: I1204 13:52:00.209641 4848 generic.go:334] "Generic (PLEG): container finished" podID="8eedf821-9e1a-4513-896f-372df7171ef2" containerID="3aef13f92912d30feed58b4eca566026cbb430ea49de87ad4bf8fc98f384a926" exitCode=0 Dec 04 13:52:00 crc kubenswrapper[4848]: I1204 13:52:00.209699 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrx58" event={"ID":"8eedf821-9e1a-4513-896f-372df7171ef2","Type":"ContainerDied","Data":"3aef13f92912d30feed58b4eca566026cbb430ea49de87ad4bf8fc98f384a926"} Dec 04 13:52:00 crc kubenswrapper[4848]: I1204 13:52:00.225331 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c2dzv" podStartSLOduration=1.793429307 podStartE2EDuration="1m8.225312361s" podCreationTimestamp="2025-12-04 13:50:52 +0000 UTC" firstStartedPulling="2025-12-04 13:50:53.580125986 +0000 UTC m=+157.522622514" lastFinishedPulling="2025-12-04 13:52:00.01200904 +0000 UTC m=+223.954505568" observedRunningTime="2025-12-04 13:52:00.222830989 +0000 UTC m=+224.165327517" watchObservedRunningTime="2025-12-04 13:52:00.225312361 +0000 UTC m=+224.167808909" Dec 04 13:52:00 crc kubenswrapper[4848]: I1204 13:52:00.412797 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75519da1-f479-431a-adbb-0713e0049bd5" path="/var/lib/kubelet/pods/75519da1-f479-431a-adbb-0713e0049bd5/volumes" Dec 04 13:52:01 crc kubenswrapper[4848]: I1204 13:52:01.222404 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5qh" event={"ID":"f3d120a8-d58e-45b9-be60-04be599332b3","Type":"ContainerStarted","Data":"f23ee88bcfb4887ef5fbbe679b4811704df7344ed0f7ddd68b391a057bef8487"} Dec 04 13:52:01 crc kubenswrapper[4848]: I1204 13:52:01.224421 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjljj" event={"ID":"e75ab632-da17-4354-951c-476ad2e46835","Type":"ContainerStarted","Data":"618cdd637d5200c2e36b73fee806765f20debe62604e7e3b00597ee7513209e9"} Dec 04 13:52:01 crc kubenswrapper[4848]: I1204 13:52:01.226663 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrx58" event={"ID":"8eedf821-9e1a-4513-896f-372df7171ef2","Type":"ContainerStarted","Data":"39abe968350d51fd71413c221f1bda289f39a61599b759f26a4669796633bc1e"} Dec 04 13:52:01 crc kubenswrapper[4848]: I1204 13:52:01.228797 4848 generic.go:334] "Generic (PLEG): container finished" podID="15f8738d-4240-42d0-a819-cccff5aaa821" containerID="9d178022a3b94d4aabe32a720b73dfedd5c341214b835068b82a7800237add6d" exitCode=0 Dec 04 13:52:01 crc kubenswrapper[4848]: I1204 13:52:01.228821 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6h7q" event={"ID":"15f8738d-4240-42d0-a819-cccff5aaa821","Type":"ContainerDied","Data":"9d178022a3b94d4aabe32a720b73dfedd5c341214b835068b82a7800237add6d"} Dec 04 13:52:01 crc kubenswrapper[4848]: I1204 13:52:01.288351 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hrx58" podStartSLOduration=3.273509698 podStartE2EDuration="1m8.288325501s" podCreationTimestamp="2025-12-04 13:50:53 +0000 UTC" firstStartedPulling="2025-12-04 13:50:55.630658546 +0000 UTC m=+159.573155074" lastFinishedPulling="2025-12-04 13:52:00.645474349 +0000 UTC m=+224.587970877" observedRunningTime="2025-12-04 13:52:01.288014754 +0000 UTC m=+225.230511282" watchObservedRunningTime="2025-12-04 13:52:01.288325501 +0000 UTC m=+225.230822039" Dec 04 13:52:02 crc kubenswrapper[4848]: I1204 13:52:02.235829 4848 generic.go:334] "Generic (PLEG): container finished" podID="f3d120a8-d58e-45b9-be60-04be599332b3" containerID="f23ee88bcfb4887ef5fbbe679b4811704df7344ed0f7ddd68b391a057bef8487" exitCode=0 Dec 04 13:52:02 crc kubenswrapper[4848]: I1204 13:52:02.235907 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5qh" event={"ID":"f3d120a8-d58e-45b9-be60-04be599332b3","Type":"ContainerDied","Data":"f23ee88bcfb4887ef5fbbe679b4811704df7344ed0f7ddd68b391a057bef8487"} Dec 04 13:52:02 crc kubenswrapper[4848]: I1204 13:52:02.240542 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjljj" event={"ID":"e75ab632-da17-4354-951c-476ad2e46835","Type":"ContainerDied","Data":"618cdd637d5200c2e36b73fee806765f20debe62604e7e3b00597ee7513209e9"} Dec 04 13:52:02 crc kubenswrapper[4848]: I1204 13:52:02.240487 4848 generic.go:334] "Generic (PLEG): container finished" podID="e75ab632-da17-4354-951c-476ad2e46835" containerID="618cdd637d5200c2e36b73fee806765f20debe62604e7e3b00597ee7513209e9" exitCode=0 Dec 04 13:52:02 crc kubenswrapper[4848]: I1204 13:52:02.742766 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:52:02 crc kubenswrapper[4848]: I1204 13:52:02.742835 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:52:02 crc kubenswrapper[4848]: I1204 13:52:02.842992 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.026586 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.066978 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.248852 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6h7q" event={"ID":"15f8738d-4240-42d0-a819-cccff5aaa821","Type":"ContainerStarted","Data":"7109be40b4f575f22c41ac3d20b08d6aa828a36c19d78c63fd064d8000571b37"} Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.252383 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5qh" event={"ID":"f3d120a8-d58e-45b9-be60-04be599332b3","Type":"ContainerStarted","Data":"3059dc72ca92d3456e72586b9cf129c03bf7347e00676d8c1483eb860d8eb816"} Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.255130 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjljj" event={"ID":"e75ab632-da17-4354-951c-476ad2e46835","Type":"ContainerStarted","Data":"0a090d37c3da4dbf63f6d6e5ade249782fd8254f8e48cd6c92e463c134cc8324"} Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.262428 4848 generic.go:334] "Generic (PLEG): container finished" podID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerID="6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046" exitCode=0 Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.263319 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r8vn4" event={"ID":"5949452c-a2a6-488a-8730-2edb4cfc5747","Type":"ContainerDied","Data":"6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046"} Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.275817 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x6h7q" podStartSLOduration=2.462126726 podStartE2EDuration="1m8.275800627s" podCreationTimestamp="2025-12-04 13:50:55 +0000 UTC" firstStartedPulling="2025-12-04 13:50:56.636559468 +0000 UTC m=+160.579055996" lastFinishedPulling="2025-12-04 13:52:02.450233369 +0000 UTC m=+226.392729897" observedRunningTime="2025-12-04 13:52:03.26990062 +0000 UTC m=+227.212397148" watchObservedRunningTime="2025-12-04 13:52:03.275800627 +0000 UTC m=+227.218297155" Dec 04 13:52:03 crc kubenswrapper[4848]: I1204 13:52:03.286750 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sjljj" podStartSLOduration=2.219497498 podStartE2EDuration="1m11.28673626s" podCreationTimestamp="2025-12-04 13:50:52 +0000 UTC" firstStartedPulling="2025-12-04 13:50:53.58921292 +0000 UTC m=+157.531709488" lastFinishedPulling="2025-12-04 13:52:02.656451722 +0000 UTC m=+226.598948250" observedRunningTime="2025-12-04 13:52:03.2831223 +0000 UTC m=+227.225618838" watchObservedRunningTime="2025-12-04 13:52:03.28673626 +0000 UTC m=+227.229232788" Dec 04 13:52:04 crc kubenswrapper[4848]: I1204 13:52:04.624085 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:52:04 crc kubenswrapper[4848]: I1204 13:52:04.624125 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:52:04 crc kubenswrapper[4848]: I1204 13:52:04.666460 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:52:04 crc kubenswrapper[4848]: I1204 13:52:04.690829 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nr5qh" podStartSLOduration=4.576089513 podStartE2EDuration="1m13.690806876s" podCreationTimestamp="2025-12-04 13:50:51 +0000 UTC" firstStartedPulling="2025-12-04 13:50:53.582733051 +0000 UTC m=+157.525229619" lastFinishedPulling="2025-12-04 13:52:02.697450454 +0000 UTC m=+226.639946982" observedRunningTime="2025-12-04 13:52:03.32280874 +0000 UTC m=+227.265305268" watchObservedRunningTime="2025-12-04 13:52:04.690806876 +0000 UTC m=+228.633303404" Dec 04 13:52:04 crc kubenswrapper[4848]: I1204 13:52:04.732058 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kp8lv"] Dec 04 13:52:04 crc kubenswrapper[4848]: I1204 13:52:04.732277 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kp8lv" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="registry-server" containerID="cri-o://28edbc3e80b44ef66937f62510bc8f12542c7f2a7f52ac597af4209b0b8a520e" gracePeriod=2 Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.274829 4848 generic.go:334] "Generic (PLEG): container finished" podID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerID="28edbc3e80b44ef66937f62510bc8f12542c7f2a7f52ac597af4209b0b8a520e" exitCode=0 Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.274889 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp8lv" event={"ID":"f98ec77d-021a-43cd-b8e9-12738f4921f1","Type":"ContainerDied","Data":"28edbc3e80b44ef66937f62510bc8f12542c7f2a7f52ac597af4209b0b8a520e"} Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.317282 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.548752 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.549161 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.724139 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.810508 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-catalog-content\") pod \"f98ec77d-021a-43cd-b8e9-12738f4921f1\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.810565 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-utilities\") pod \"f98ec77d-021a-43cd-b8e9-12738f4921f1\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.810655 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkqg2\" (UniqueName: \"kubernetes.io/projected/f98ec77d-021a-43cd-b8e9-12738f4921f1-kube-api-access-mkqg2\") pod \"f98ec77d-021a-43cd-b8e9-12738f4921f1\" (UID: \"f98ec77d-021a-43cd-b8e9-12738f4921f1\") " Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.811466 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-utilities" (OuterVolumeSpecName: "utilities") pod "f98ec77d-021a-43cd-b8e9-12738f4921f1" (UID: "f98ec77d-021a-43cd-b8e9-12738f4921f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.821119 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f98ec77d-021a-43cd-b8e9-12738f4921f1-kube-api-access-mkqg2" (OuterVolumeSpecName: "kube-api-access-mkqg2") pod "f98ec77d-021a-43cd-b8e9-12738f4921f1" (UID: "f98ec77d-021a-43cd-b8e9-12738f4921f1"). InnerVolumeSpecName "kube-api-access-mkqg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.863474 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f98ec77d-021a-43cd-b8e9-12738f4921f1" (UID: "f98ec77d-021a-43cd-b8e9-12738f4921f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.912866 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.912935 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkqg2\" (UniqueName: \"kubernetes.io/projected/f98ec77d-021a-43cd-b8e9-12738f4921f1-kube-api-access-mkqg2\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:05 crc kubenswrapper[4848]: I1204 13:52:05.912969 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98ec77d-021a-43cd-b8e9-12738f4921f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.283130 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r8vn4" event={"ID":"5949452c-a2a6-488a-8730-2edb4cfc5747","Type":"ContainerStarted","Data":"51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0"} Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.285778 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kp8lv" Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.289430 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kp8lv" event={"ID":"f98ec77d-021a-43cd-b8e9-12738f4921f1","Type":"ContainerDied","Data":"aedbbf795cced2e3ea02cd5698326139e2168d2be173963a6e5f406c3975baf1"} Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.289473 4848 scope.go:117] "RemoveContainer" containerID="28edbc3e80b44ef66937f62510bc8f12542c7f2a7f52ac597af4209b0b8a520e" Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.302230 4848 scope.go:117] "RemoveContainer" containerID="47893295eba47e208843c9bd972e6af31a4fa0f3eb7edfa12f35fd970cd36da8" Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.316548 4848 scope.go:117] "RemoveContainer" containerID="4635ad052766c3821eb228a283ed7226b1460d0f16d8fcc1c1598b89af25e286" Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.348277 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kp8lv"] Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.350676 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kp8lv"] Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.398851 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" path="/var/lib/kubelet/pods/f98ec77d-021a-43cd-b8e9-12738f4921f1/volumes" Dec 04 13:52:06 crc kubenswrapper[4848]: I1204 13:52:06.608437 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x6h7q" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="registry-server" probeResult="failure" output=< Dec 04 13:52:06 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 13:52:06 crc kubenswrapper[4848]: > Dec 04 13:52:07 crc kubenswrapper[4848]: I1204 13:52:07.311330 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r8vn4" podStartSLOduration=3.523526597 podStartE2EDuration="1m12.31130677s" podCreationTimestamp="2025-12-04 13:50:55 +0000 UTC" firstStartedPulling="2025-12-04 13:50:56.633325847 +0000 UTC m=+160.575822365" lastFinishedPulling="2025-12-04 13:52:05.42110601 +0000 UTC m=+229.363602538" observedRunningTime="2025-12-04 13:52:07.307328591 +0000 UTC m=+231.249825159" watchObservedRunningTime="2025-12-04 13:52:07.31130677 +0000 UTC m=+231.253803318" Dec 04 13:52:12 crc kubenswrapper[4848]: I1204 13:52:12.331995 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:52:12 crc kubenswrapper[4848]: I1204 13:52:12.332280 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:52:12 crc kubenswrapper[4848]: I1204 13:52:12.369577 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:52:12 crc kubenswrapper[4848]: I1204 13:52:12.610815 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:52:12 crc kubenswrapper[4848]: I1204 13:52:12.610868 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:52:12 crc kubenswrapper[4848]: I1204 13:52:12.674214 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:52:12 crc kubenswrapper[4848]: I1204 13:52:12.779238 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:52:13 crc kubenswrapper[4848]: I1204 13:52:13.380618 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:52:13 crc kubenswrapper[4848]: I1204 13:52:13.384361 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:52:13 crc kubenswrapper[4848]: I1204 13:52:13.609995 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c2dzv"] Dec 04 13:52:13 crc kubenswrapper[4848]: I1204 13:52:13.610252 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c2dzv" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerName="registry-server" containerID="cri-o://a2543378f8bc40ef105c55cd3433908fbf6f08574737dd98bf4ea4c833c9a086" gracePeriod=2 Dec 04 13:52:15 crc kubenswrapper[4848]: I1204 13:52:15.600548 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:52:15 crc kubenswrapper[4848]: I1204 13:52:15.659449 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:52:15 crc kubenswrapper[4848]: I1204 13:52:15.967349 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:52:15 crc kubenswrapper[4848]: I1204 13:52:15.967407 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:52:16 crc kubenswrapper[4848]: I1204 13:52:16.040539 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:52:16 crc kubenswrapper[4848]: I1204 13:52:16.387219 4848 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-lwll8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 13:52:16 crc kubenswrapper[4848]: I1204 13:52:16.387304 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" podUID="82ae395f-c4ee-43de-a1d6-0349c6dc640b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 13:52:16 crc kubenswrapper[4848]: I1204 13:52:16.416152 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:52:19 crc kubenswrapper[4848]: I1204 13:52:19.791939 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dmbbx"] Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.154558 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.196648 4848 generic.go:334] "Generic (PLEG): container finished" podID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerID="a2543378f8bc40ef105c55cd3433908fbf6f08574737dd98bf4ea4c833c9a086" exitCode=0 Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.196689 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2dzv" event={"ID":"00f2c84d-1e46-4563-9c78-6657bb601db3","Type":"ContainerDied","Data":"a2543378f8bc40ef105c55cd3433908fbf6f08574737dd98bf4ea4c833c9a086"} Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.196722 4848 scope.go:117] "RemoveContainer" containerID="a2543378f8bc40ef105c55cd3433908fbf6f08574737dd98bf4ea4c833c9a086" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.214611 4848 scope.go:117] "RemoveContainer" containerID="2c1fb7a81aca3ff4e2b951c018a99f7ce6e0056646fc8aae4c3dce2b64bfc35d" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.229914 4848 scope.go:117] "RemoveContainer" containerID="8d74ab55ac61323e251f69651a9514773409356a02c83211c6f7d7f8f433d649" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.294197 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-catalog-content\") pod \"00f2c84d-1e46-4563-9c78-6657bb601db3\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.294285 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-utilities\") pod \"00f2c84d-1e46-4563-9c78-6657bb601db3\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.294381 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfblq\" (UniqueName: \"kubernetes.io/projected/00f2c84d-1e46-4563-9c78-6657bb601db3-kube-api-access-wfblq\") pod \"00f2c84d-1e46-4563-9c78-6657bb601db3\" (UID: \"00f2c84d-1e46-4563-9c78-6657bb601db3\") " Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.295494 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-utilities" (OuterVolumeSpecName: "utilities") pod "00f2c84d-1e46-4563-9c78-6657bb601db3" (UID: "00f2c84d-1e46-4563-9c78-6657bb601db3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.300290 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f2c84d-1e46-4563-9c78-6657bb601db3-kube-api-access-wfblq" (OuterVolumeSpecName: "kube-api-access-wfblq") pod "00f2c84d-1e46-4563-9c78-6657bb601db3" (UID: "00f2c84d-1e46-4563-9c78-6657bb601db3"). InnerVolumeSpecName "kube-api-access-wfblq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.355515 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00f2c84d-1e46-4563-9c78-6657bb601db3" (UID: "00f2c84d-1e46-4563-9c78-6657bb601db3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.395584 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.395618 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00f2c84d-1e46-4563-9c78-6657bb601db3-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:20 crc kubenswrapper[4848]: I1204 13:52:20.395632 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfblq\" (UniqueName: \"kubernetes.io/projected/00f2c84d-1e46-4563-9c78-6657bb601db3-kube-api-access-wfblq\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:21 crc kubenswrapper[4848]: I1204 13:52:21.202419 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c2dzv" event={"ID":"00f2c84d-1e46-4563-9c78-6657bb601db3","Type":"ContainerDied","Data":"1b7d3a6d808b16283114a935c37f75e7abdc9d9af23ec2f3ca291d1f9e377590"} Dec 04 13:52:21 crc kubenswrapper[4848]: I1204 13:52:21.202532 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c2dzv" Dec 04 13:52:21 crc kubenswrapper[4848]: I1204 13:52:21.221439 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c2dzv"] Dec 04 13:52:21 crc kubenswrapper[4848]: I1204 13:52:21.232329 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c2dzv"] Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.018791 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r8vn4"] Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.019201 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r8vn4" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerName="registry-server" containerID="cri-o://51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0" gracePeriod=2 Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.407061 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" path="/var/lib/kubelet/pods/00f2c84d-1e46-4563-9c78-6657bb601db3/volumes" Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.925107 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.935328 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-catalog-content\") pod \"5949452c-a2a6-488a-8730-2edb4cfc5747\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.935411 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnzzw\" (UniqueName: \"kubernetes.io/projected/5949452c-a2a6-488a-8730-2edb4cfc5747-kube-api-access-xnzzw\") pod \"5949452c-a2a6-488a-8730-2edb4cfc5747\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.935554 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-utilities\") pod \"5949452c-a2a6-488a-8730-2edb4cfc5747\" (UID: \"5949452c-a2a6-488a-8730-2edb4cfc5747\") " Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.936798 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-utilities" (OuterVolumeSpecName: "utilities") pod "5949452c-a2a6-488a-8730-2edb4cfc5747" (UID: "5949452c-a2a6-488a-8730-2edb4cfc5747"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:52:22 crc kubenswrapper[4848]: I1204 13:52:22.944357 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5949452c-a2a6-488a-8730-2edb4cfc5747-kube-api-access-xnzzw" (OuterVolumeSpecName: "kube-api-access-xnzzw") pod "5949452c-a2a6-488a-8730-2edb4cfc5747" (UID: "5949452c-a2a6-488a-8730-2edb4cfc5747"). InnerVolumeSpecName "kube-api-access-xnzzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.036216 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.036257 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnzzw\" (UniqueName: \"kubernetes.io/projected/5949452c-a2a6-488a-8730-2edb4cfc5747-kube-api-access-xnzzw\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.067169 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5949452c-a2a6-488a-8730-2edb4cfc5747" (UID: "5949452c-a2a6-488a-8730-2edb4cfc5747"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.137339 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5949452c-a2a6-488a-8730-2edb4cfc5747-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.217607 4848 generic.go:334] "Generic (PLEG): container finished" podID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerID="51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0" exitCode=0 Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.217649 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r8vn4" event={"ID":"5949452c-a2a6-488a-8730-2edb4cfc5747","Type":"ContainerDied","Data":"51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0"} Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.217671 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r8vn4" event={"ID":"5949452c-a2a6-488a-8730-2edb4cfc5747","Type":"ContainerDied","Data":"22f29693a6454a29ae4effcaf52be2d69834f612e9e3660c4d2add5df9709ee4"} Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.217687 4848 scope.go:117] "RemoveContainer" containerID="51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.217715 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r8vn4" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.243255 4848 scope.go:117] "RemoveContainer" containerID="6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.256909 4848 scope.go:117] "RemoveContainer" containerID="b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.265399 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r8vn4"] Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.269032 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r8vn4"] Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.281076 4848 scope.go:117] "RemoveContainer" containerID="51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0" Dec 04 13:52:23 crc kubenswrapper[4848]: E1204 13:52:23.281419 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0\": container with ID starting with 51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0 not found: ID does not exist" containerID="51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.281456 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0"} err="failed to get container status \"51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0\": rpc error: code = NotFound desc = could not find container \"51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0\": container with ID starting with 51713d20822ac728027da6d3d985eb26df577e754429be83b33b2e200d03dfa0 not found: ID does not exist" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.281478 4848 scope.go:117] "RemoveContainer" containerID="6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046" Dec 04 13:52:23 crc kubenswrapper[4848]: E1204 13:52:23.281704 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046\": container with ID starting with 6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046 not found: ID does not exist" containerID="6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.281726 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046"} err="failed to get container status \"6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046\": rpc error: code = NotFound desc = could not find container \"6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046\": container with ID starting with 6d5210f8a549937c0efed4e8715957cdc8e97b121b718e2cfe6db7ee3fdca046 not found: ID does not exist" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.281739 4848 scope.go:117] "RemoveContainer" containerID="b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac" Dec 04 13:52:23 crc kubenswrapper[4848]: E1204 13:52:23.281916 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac\": container with ID starting with b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac not found: ID does not exist" containerID="b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac" Dec 04 13:52:23 crc kubenswrapper[4848]: I1204 13:52:23.281932 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac"} err="failed to get container status \"b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac\": rpc error: code = NotFound desc = could not find container \"b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac\": container with ID starting with b4e826183556a591b924e6fce43b0c7c169940bacac50fab97259eed4649a9ac not found: ID does not exist" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.403137 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" path="/var/lib/kubelet/pods/5949452c-a2a6-488a-8730-2edb4cfc5747/volumes" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.776524 4848 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.776799 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf" gracePeriod=15 Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.776907 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803" gracePeriod=15 Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.776911 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b" gracePeriod=15 Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.777030 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5" gracePeriod=15 Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.777368 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82" gracePeriod=15 Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.778788 4848 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782117 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782159 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782182 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerName="extract-content" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782195 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerName="extract-content" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782217 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerName="extract-utilities" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782231 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerName="extract-utilities" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782247 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782260 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782273 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782286 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782306 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782318 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782336 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerName="extract-content" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782351 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerName="extract-content" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782372 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782385 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782403 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60783e1c-df42-4e49-90d9-1aa0a1fef050" containerName="pruner" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782417 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="60783e1c-df42-4e49-90d9-1aa0a1fef050" containerName="pruner" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782431 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782444 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782464 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="extract-content" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782477 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="extract-content" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782497 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="extract-utilities" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782510 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="extract-utilities" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782533 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75519da1-f479-431a-adbb-0713e0049bd5" containerName="extract-content" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782546 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="75519da1-f479-431a-adbb-0713e0049bd5" containerName="extract-content" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782564 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782578 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782606 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782618 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782635 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerName="extract-utilities" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782650 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerName="extract-utilities" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782674 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75519da1-f479-431a-adbb-0713e0049bd5" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782691 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="75519da1-f479-431a-adbb-0713e0049bd5" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782711 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782724 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.782753 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75519da1-f479-431a-adbb-0713e0049bd5" containerName="extract-utilities" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.782770 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="75519da1-f479-431a-adbb-0713e0049bd5" containerName="extract-utilities" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783069 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783100 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783124 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783142 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783155 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="75519da1-f479-431a-adbb-0713e0049bd5" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783175 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f98ec77d-021a-43cd-b8e9-12738f4921f1" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783190 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5949452c-a2a6-488a-8730-2edb4cfc5747" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783206 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f2c84d-1e46-4563-9c78-6657bb601db3" containerName="registry-server" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783221 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="60783e1c-df42-4e49-90d9-1aa0a1fef050" containerName="pruner" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.783237 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.792146 4848 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.792815 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.798541 4848 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 04 13:52:24 crc kubenswrapper[4848]: E1204 13:52:24.842611 4848 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.5:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.959281 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.959361 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.959433 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.959461 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.959505 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.959532 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.959625 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:24 crc kubenswrapper[4848]: I1204 13:52:24.959675 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060345 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060420 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060448 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060467 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060497 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060526 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060528 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060560 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060590 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060604 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060581 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060622 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060644 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060652 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060671 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.060842 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.143586 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:25 crc kubenswrapper[4848]: W1204 13:52:25.165458 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-800bdbb020becf8302a3adad36a3ebf0f603c7d1e2ebb68612b53722e9384d25 WatchSource:0}: Error finding container 800bdbb020becf8302a3adad36a3ebf0f603c7d1e2ebb68612b53722e9384d25: Status 404 returned error can't find the container with id 800bdbb020becf8302a3adad36a3ebf0f603c7d1e2ebb68612b53722e9384d25 Dec 04 13:52:25 crc kubenswrapper[4848]: E1204 13:52:25.168545 4848 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.5:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e077ecd4fa371 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 13:52:25.167635313 +0000 UTC m=+249.110131881,LastTimestamp:2025-12-04 13:52:25.167635313 +0000 UTC m=+249.110131881,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.232917 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.234473 4848 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b" exitCode=0 Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.234512 4848 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5" exitCode=0 Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.234526 4848 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82" exitCode=0 Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.234535 4848 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803" exitCode=2 Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.235918 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"800bdbb020becf8302a3adad36a3ebf0f603c7d1e2ebb68612b53722e9384d25"} Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.237506 4848 generic.go:334] "Generic (PLEG): container finished" podID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" containerID="2382242c375993091efabeae4815a7c23bc6d65e947f25d730662d6dba6ba24c" exitCode=0 Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.237538 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0a33ea42-e76c-4853-b7ad-68f96192d8f4","Type":"ContainerDied","Data":"2382242c375993091efabeae4815a7c23bc6d65e947f25d730662d6dba6ba24c"} Dec 04 13:52:25 crc kubenswrapper[4848]: I1204 13:52:25.238142 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.244209 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088"} Dec 04 13:52:26 crc kubenswrapper[4848]: E1204 13:52:26.244887 4848 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.5:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.244881 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.398680 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.465465 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.466089 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.578151 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kube-api-access\") pod \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.578224 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kubelet-dir\") pod \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.578289 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-var-lock\") pod \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\" (UID: \"0a33ea42-e76c-4853-b7ad-68f96192d8f4\") " Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.578332 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0a33ea42-e76c-4853-b7ad-68f96192d8f4" (UID: "0a33ea42-e76c-4853-b7ad-68f96192d8f4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.578436 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-var-lock" (OuterVolumeSpecName: "var-lock") pod "0a33ea42-e76c-4853-b7ad-68f96192d8f4" (UID: "0a33ea42-e76c-4853-b7ad-68f96192d8f4"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.578721 4848 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.578747 4848 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.583903 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0a33ea42-e76c-4853-b7ad-68f96192d8f4" (UID: "0a33ea42-e76c-4853-b7ad-68f96192d8f4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:52:26 crc kubenswrapper[4848]: I1204 13:52:26.680254 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a33ea42-e76c-4853-b7ad-68f96192d8f4-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.140160 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.141501 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.142232 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.143061 4848 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.254414 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.255308 4848 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf" exitCode=0 Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.255413 4848 scope.go:117] "RemoveContainer" containerID="24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.255520 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.259120 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"0a33ea42-e76c-4853-b7ad-68f96192d8f4","Type":"ContainerDied","Data":"e97e006ea342c1cb897f9b4d53e4ed71b80a99b7f76ea5140d5a7288b93aca90"} Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.259175 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e97e006ea342c1cb897f9b4d53e4ed71b80a99b7f76ea5140d5a7288b93aca90" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.259184 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 13:52:27 crc kubenswrapper[4848]: E1204 13:52:27.259268 4848 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.5:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.280032 4848 scope.go:117] "RemoveContainer" containerID="2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.289748 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.289870 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.289926 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.289923 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.289971 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.290002 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.291206 4848 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.291224 4848 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.291235 4848 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.296086 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.296755 4848 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.304677 4848 scope.go:117] "RemoveContainer" containerID="762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.321829 4848 scope.go:117] "RemoveContainer" containerID="c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.338380 4848 scope.go:117] "RemoveContainer" containerID="265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.351537 4848 scope.go:117] "RemoveContainer" containerID="9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.364479 4848 scope.go:117] "RemoveContainer" containerID="24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b" Dec 04 13:52:27 crc kubenswrapper[4848]: E1204 13:52:27.364921 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\": container with ID starting with 24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b not found: ID does not exist" containerID="24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.364984 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b"} err="failed to get container status \"24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\": rpc error: code = NotFound desc = could not find container \"24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b\": container with ID starting with 24b564670216897aa24cad91e3accac3d2a0466a842a990a29b8e0ca38c2f67b not found: ID does not exist" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.365015 4848 scope.go:117] "RemoveContainer" containerID="2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5" Dec 04 13:52:27 crc kubenswrapper[4848]: E1204 13:52:27.365346 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\": container with ID starting with 2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5 not found: ID does not exist" containerID="2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.365387 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5"} err="failed to get container status \"2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\": rpc error: code = NotFound desc = could not find container \"2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5\": container with ID starting with 2ae9d9cb6318c0b84c42fc735d099aaf4338376514a22828284d2bd86bd9d1a5 not found: ID does not exist" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.365419 4848 scope.go:117] "RemoveContainer" containerID="762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82" Dec 04 13:52:27 crc kubenswrapper[4848]: E1204 13:52:27.365758 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\": container with ID starting with 762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82 not found: ID does not exist" containerID="762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.365797 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82"} err="failed to get container status \"762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\": rpc error: code = NotFound desc = could not find container \"762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82\": container with ID starting with 762bab3407c7920712a6aa71242273a5435e9b8f36c0fa9038e21a7a04404a82 not found: ID does not exist" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.365819 4848 scope.go:117] "RemoveContainer" containerID="c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803" Dec 04 13:52:27 crc kubenswrapper[4848]: E1204 13:52:27.366512 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\": container with ID starting with c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803 not found: ID does not exist" containerID="c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.366540 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803"} err="failed to get container status \"c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\": rpc error: code = NotFound desc = could not find container \"c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803\": container with ID starting with c43571a7305b8649d7bd859911304589859d1eb6fdb005599d217ca8d1b51803 not found: ID does not exist" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.366560 4848 scope.go:117] "RemoveContainer" containerID="265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf" Dec 04 13:52:27 crc kubenswrapper[4848]: E1204 13:52:27.366903 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\": container with ID starting with 265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf not found: ID does not exist" containerID="265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.366933 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf"} err="failed to get container status \"265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\": rpc error: code = NotFound desc = could not find container \"265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf\": container with ID starting with 265d33b7c5ceeeb0bb0e2d78452f6fff1b89cc20ed152670c1c478923ca87acf not found: ID does not exist" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.366976 4848 scope.go:117] "RemoveContainer" containerID="9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6" Dec 04 13:52:27 crc kubenswrapper[4848]: E1204 13:52:27.367516 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\": container with ID starting with 9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6 not found: ID does not exist" containerID="9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.367548 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6"} err="failed to get container status \"9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\": rpc error: code = NotFound desc = could not find container \"9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6\": container with ID starting with 9ee542547461c9e9e4942f826fe1778bc534453d2d2075552f9f164aaba675f6 not found: ID does not exist" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.571369 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:27 crc kubenswrapper[4848]: I1204 13:52:27.571889 4848 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:28 crc kubenswrapper[4848]: I1204 13:52:28.409053 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 04 13:52:29 crc kubenswrapper[4848]: E1204 13:52:29.417349 4848 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.5:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" volumeName="registry-storage" Dec 04 13:52:31 crc kubenswrapper[4848]: E1204 13:52:31.274915 4848 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:31 crc kubenswrapper[4848]: E1204 13:52:31.275741 4848 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:31 crc kubenswrapper[4848]: E1204 13:52:31.276129 4848 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:31 crc kubenswrapper[4848]: E1204 13:52:31.276405 4848 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:31 crc kubenswrapper[4848]: E1204 13:52:31.276683 4848 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:31 crc kubenswrapper[4848]: I1204 13:52:31.276725 4848 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 04 13:52:31 crc kubenswrapper[4848]: E1204 13:52:31.277079 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="200ms" Dec 04 13:52:31 crc kubenswrapper[4848]: E1204 13:52:31.477802 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="400ms" Dec 04 13:52:31 crc kubenswrapper[4848]: E1204 13:52:31.879407 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="800ms" Dec 04 13:52:32 crc kubenswrapper[4848]: E1204 13:52:32.680639 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="1.6s" Dec 04 13:52:33 crc kubenswrapper[4848]: E1204 13:52:33.763722 4848 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.5:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e077ecd4fa371 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 13:52:25.167635313 +0000 UTC m=+249.110131881,LastTimestamp:2025-12-04 13:52:25.167635313 +0000 UTC m=+249.110131881,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 13:52:34 crc kubenswrapper[4848]: E1204 13:52:34.283071 4848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.5:6443: connect: connection refused" interval="3.2s" Dec 04 13:52:35 crc kubenswrapper[4848]: I1204 13:52:35.397347 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:35 crc kubenswrapper[4848]: I1204 13:52:35.399103 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:35 crc kubenswrapper[4848]: I1204 13:52:35.420826 4848 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d58b6feb-e01f-4631-bce7-26eb2f6aea37" Dec 04 13:52:35 crc kubenswrapper[4848]: I1204 13:52:35.420873 4848 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d58b6feb-e01f-4631-bce7-26eb2f6aea37" Dec 04 13:52:35 crc kubenswrapper[4848]: E1204 13:52:35.421462 4848 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:35 crc kubenswrapper[4848]: I1204 13:52:35.422093 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:36 crc kubenswrapper[4848]: I1204 13:52:36.313284 4848 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d546796a4ca47dfaca982cc2d7152506d4c732cce8a37b129d6d0283ca9288dc" exitCode=0 Dec 04 13:52:36 crc kubenswrapper[4848]: I1204 13:52:36.313376 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d546796a4ca47dfaca982cc2d7152506d4c732cce8a37b129d6d0283ca9288dc"} Dec 04 13:52:36 crc kubenswrapper[4848]: I1204 13:52:36.313756 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f9293645c3f35cd21145a454df49fa809b7242cdabe02d4046fa0169dc0d03b0"} Dec 04 13:52:36 crc kubenswrapper[4848]: I1204 13:52:36.314310 4848 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d58b6feb-e01f-4631-bce7-26eb2f6aea37" Dec 04 13:52:36 crc kubenswrapper[4848]: I1204 13:52:36.314355 4848 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d58b6feb-e01f-4631-bce7-26eb2f6aea37" Dec 04 13:52:36 crc kubenswrapper[4848]: I1204 13:52:36.314897 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:36 crc kubenswrapper[4848]: E1204 13:52:36.315242 4848 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:36 crc kubenswrapper[4848]: I1204 13:52:36.405796 4848 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:36 crc kubenswrapper[4848]: I1204 13:52:36.406406 4848 status_manager.go:851] "Failed to get status for pod" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.5:6443: connect: connection refused" Dec 04 13:52:37 crc kubenswrapper[4848]: I1204 13:52:37.325124 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4d6d880fa31c08d9a62405e93f06c63273c20fb18538e3d3ee409aa41e7d032e"} Dec 04 13:52:37 crc kubenswrapper[4848]: I1204 13:52:37.325422 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"97c4c0eacd3bd3de903601bb703eeee914d3f4776bed79a39bf17a434dd9cb56"} Dec 04 13:52:37 crc kubenswrapper[4848]: I1204 13:52:37.325433 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8e226b2b9e3a8872bdff2cc3ebf3e3cc0080ca35eaa3b5d5d75cb4fe7ae6ecd1"} Dec 04 13:52:37 crc kubenswrapper[4848]: I1204 13:52:37.325442 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"97f2c04c2b469e6d37a819edbc2cd28cc9f384de165d0563aa9f17f63e306b74"} Dec 04 13:52:38 crc kubenswrapper[4848]: I1204 13:52:38.332637 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3e21e2b7ff15650bbbb934362fab7bbd6404ebf9c44d2ebd0f0dc0eb58e44296"} Dec 04 13:52:38 crc kubenswrapper[4848]: I1204 13:52:38.332975 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:38 crc kubenswrapper[4848]: I1204 13:52:38.333091 4848 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d58b6feb-e01f-4631-bce7-26eb2f6aea37" Dec 04 13:52:38 crc kubenswrapper[4848]: I1204 13:52:38.333118 4848 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d58b6feb-e01f-4631-bce7-26eb2f6aea37" Dec 04 13:52:39 crc kubenswrapper[4848]: I1204 13:52:39.610082 4848 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 04 13:52:39 crc kubenswrapper[4848]: I1204 13:52:39.610201 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 04 13:52:40 crc kubenswrapper[4848]: I1204 13:52:40.398516 4848 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 04 13:52:40 crc kubenswrapper[4848]: I1204 13:52:40.398855 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 04 13:52:40 crc kubenswrapper[4848]: I1204 13:52:40.422329 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:40 crc kubenswrapper[4848]: I1204 13:52:40.422394 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:40 crc kubenswrapper[4848]: I1204 13:52:40.431536 4848 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]log ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]etcd ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/generic-apiserver-start-informers ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/priority-and-fairness-filter ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-apiextensions-informers ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-apiextensions-controllers ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/crd-informer-synced ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-system-namespaces-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 04 13:52:40 crc kubenswrapper[4848]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/bootstrap-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/start-kube-aggregator-informers ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/apiservice-registration-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/apiservice-discovery-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]autoregister-completion ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/apiservice-openapi-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 04 13:52:40 crc kubenswrapper[4848]: livez check failed Dec 04 13:52:40 crc kubenswrapper[4848]: I1204 13:52:40.431632 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 13:52:42 crc kubenswrapper[4848]: I1204 13:52:42.358926 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 13:52:42 crc kubenswrapper[4848]: I1204 13:52:42.358993 4848 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66" exitCode=1 Dec 04 13:52:42 crc kubenswrapper[4848]: I1204 13:52:42.359070 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66"} Dec 04 13:52:42 crc kubenswrapper[4848]: I1204 13:52:42.359545 4848 scope.go:117] "RemoveContainer" containerID="d5436673695072f02eb1808c1baa91e1ff6d70aa09b6a6d8f2041160c7b99c66" Dec 04 13:52:43 crc kubenswrapper[4848]: I1204 13:52:43.347543 4848 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:52:43 crc kubenswrapper[4848]: I1204 13:52:43.367844 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 13:52:43 crc kubenswrapper[4848]: I1204 13:52:43.367893 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"708f07d144b8b611c3b67f649bb352f33c148ddf709eff11bf934beaf843e0d9"} Dec 04 13:52:43 crc kubenswrapper[4848]: I1204 13:52:43.371896 4848 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58b6feb-e01f-4631-bce7-26eb2f6aea37\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:52:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:52:36Z\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T13:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://97f2c04c2b469e6d37a819edbc2cd28cc9f384de165d0563aa9f17f63e306b74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://97c4c0eacd3bd3de903601bb703eeee914d3f4776bed79a39bf17a434dd9cb56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e226b2b9e3a8872bdff2cc3ebf3e3cc0080ca35eaa3b5d5d75cb4fe7ae6ecd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e21e2b7ff15650bbbb934362fab7bbd6404ebf9c44d2ebd0f0dc0eb58e44296\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d6d880fa31c08d9a62405e93f06c63273c20fb18538e3d3ee409aa41e7d032e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T13:52:37Z\\\"}}}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d546796a4ca47dfaca982cc2d7152506d4c732cce8a37b129d6d0283ca9288dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d546796a4ca47dfaca982cc2d7152506d4c732cce8a37b129d6d0283ca9288dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T13:52:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T13:52:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}]}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Pod \"kube-apiserver-crc\" is invalid: metadata.uid: Invalid value: \"d58b6feb-e01f-4631-bce7-26eb2f6aea37\": field is immutable" Dec 04 13:52:43 crc kubenswrapper[4848]: I1204 13:52:43.389332 4848 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b58fab04-f28c-45ee-8d5f-687c4aa56cb9" Dec 04 13:52:44 crc kubenswrapper[4848]: I1204 13:52:44.374834 4848 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d58b6feb-e01f-4631-bce7-26eb2f6aea37" Dec 04 13:52:44 crc kubenswrapper[4848]: I1204 13:52:44.375388 4848 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d58b6feb-e01f-4631-bce7-26eb2f6aea37" Dec 04 13:52:44 crc kubenswrapper[4848]: I1204 13:52:44.378691 4848 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b58fab04-f28c-45ee-8d5f-687c4aa56cb9" Dec 04 13:52:44 crc kubenswrapper[4848]: I1204 13:52:44.831064 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" podUID="eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" containerName="oauth-openshift" containerID="cri-o://616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7" gracePeriod=15 Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.260438 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.380634 4848 generic.go:334] "Generic (PLEG): container finished" podID="eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" containerID="616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7" exitCode=0 Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.380678 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" event={"ID":"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c","Type":"ContainerDied","Data":"616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7"} Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.380703 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" event={"ID":"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c","Type":"ContainerDied","Data":"3da91df77c615951cb70d76df6e4a249d18aa20f365185064725042944bbecbc"} Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.380724 4848 scope.go:117] "RemoveContainer" containerID="616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.380891 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dmbbx" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.411832 4848 scope.go:117] "RemoveContainer" containerID="616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7" Dec 04 13:52:45 crc kubenswrapper[4848]: E1204 13:52:45.412425 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7\": container with ID starting with 616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7 not found: ID does not exist" containerID="616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.412484 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7"} err="failed to get container status \"616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7\": rpc error: code = NotFound desc = could not find container \"616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7\": container with ID starting with 616c58ff900789aa0922c02aaa7d70f79a72c8b170190085328955ddd5976ab7 not found: ID does not exist" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425403 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnv8h\" (UniqueName: \"kubernetes.io/projected/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-kube-api-access-hnv8h\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425472 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-error\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425498 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-session\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425522 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-dir\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425560 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-login\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425614 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-provider-selection\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425650 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-cliconfig\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425682 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-serving-cert\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425698 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425717 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-trusted-ca-bundle\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425805 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-service-ca\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425871 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-ocp-branding-template\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.425981 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-policies\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.426051 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-router-certs\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.426091 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-idp-0-file-data\") pod \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\" (UID: \"eefea8cf-d6b0-45bf-8be8-1abf545d0e5c\") " Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.426473 4848 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.427018 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.427581 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.431074 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.432433 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-kube-api-access-hnv8h" (OuterVolumeSpecName: "kube-api-access-hnv8h") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "kube-api-access-hnv8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.432902 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.433377 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.433635 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.434266 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.434573 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.440094 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.443256 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.444152 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.444881 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" (UID: "eefea8cf-d6b0-45bf-8be8-1abf545d0e5c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527848 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnv8h\" (UniqueName: \"kubernetes.io/projected/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-kube-api-access-hnv8h\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527884 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527902 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527916 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527931 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527961 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527974 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527987 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.527999 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.528013 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.528026 4848 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.528038 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:45 crc kubenswrapper[4848]: I1204 13:52:45.528051 4848 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:52:46 crc kubenswrapper[4848]: E1204 13:52:46.902365 4848 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Dec 04 13:52:48 crc kubenswrapper[4848]: I1204 13:52:48.355072 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:52:48 crc kubenswrapper[4848]: I1204 13:52:48.355380 4848 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 04 13:52:48 crc kubenswrapper[4848]: I1204 13:52:48.355441 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 04 13:52:49 crc kubenswrapper[4848]: I1204 13:52:49.530446 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 13:52:49 crc kubenswrapper[4848]: I1204 13:52:49.860290 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 13:52:49 crc kubenswrapper[4848]: I1204 13:52:49.980729 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 13:52:50 crc kubenswrapper[4848]: I1204 13:52:50.407151 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:52:50 crc kubenswrapper[4848]: I1204 13:52:50.692428 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 13:52:50 crc kubenswrapper[4848]: I1204 13:52:50.957414 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 13:52:51 crc kubenswrapper[4848]: I1204 13:52:51.562153 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 13:52:51 crc kubenswrapper[4848]: I1204 13:52:51.888668 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 13:52:51 crc kubenswrapper[4848]: I1204 13:52:51.940786 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 13:52:52 crc kubenswrapper[4848]: I1204 13:52:52.715371 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 13:52:52 crc kubenswrapper[4848]: I1204 13:52:52.974401 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 13:52:53 crc kubenswrapper[4848]: I1204 13:52:53.270435 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 13:52:53 crc kubenswrapper[4848]: I1204 13:52:53.674896 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 13:52:54 crc kubenswrapper[4848]: I1204 13:52:54.201391 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 13:52:54 crc kubenswrapper[4848]: I1204 13:52:54.400509 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 13:52:54 crc kubenswrapper[4848]: I1204 13:52:54.426396 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 13:52:54 crc kubenswrapper[4848]: I1204 13:52:54.470783 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 13:52:54 crc kubenswrapper[4848]: I1204 13:52:54.585397 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 13:52:55 crc kubenswrapper[4848]: I1204 13:52:55.020837 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 13:52:55 crc kubenswrapper[4848]: I1204 13:52:55.232589 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 13:52:55 crc kubenswrapper[4848]: I1204 13:52:55.504293 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 13:52:55 crc kubenswrapper[4848]: I1204 13:52:55.965344 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 13:52:56 crc kubenswrapper[4848]: I1204 13:52:56.327117 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 13:52:56 crc kubenswrapper[4848]: I1204 13:52:56.368653 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 13:52:56 crc kubenswrapper[4848]: I1204 13:52:56.386989 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 13:52:56 crc kubenswrapper[4848]: I1204 13:52:56.988919 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.077712 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.159552 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.209419 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.426714 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.427863 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.655614 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.871160 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.931857 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 13:52:57 crc kubenswrapper[4848]: I1204 13:52:57.932019 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.030404 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.335196 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.365197 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.375297 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.504002 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.597777 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.643589 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.766929 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.808018 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 13:52:58 crc kubenswrapper[4848]: I1204 13:52:58.828609 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.058527 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.060190 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.096646 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.554014 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.619568 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.693723 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.705736 4848 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.749697 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.791446 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.821686 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 13:52:59 crc kubenswrapper[4848]: I1204 13:52:59.990430 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.033817 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.061736 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.129761 4848 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.160099 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.249777 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.512980 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.552272 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.581274 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.791595 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.896076 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.911293 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 13:53:00 crc kubenswrapper[4848]: I1204 13:53:00.986680 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.172072 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.199450 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.199581 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.225613 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.307275 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.322174 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.377269 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.398763 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.409732 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.430788 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.581310 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.600847 4848 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.715097 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.889045 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.889066 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 13:53:01 crc kubenswrapper[4848]: I1204 13:53:01.902300 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.006200 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.063664 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.134407 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.153065 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.291141 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.353356 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.437199 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.459482 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.498398 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.516041 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.630562 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.674269 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.696792 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.811644 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.820765 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.838624 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 13:53:02 crc kubenswrapper[4848]: I1204 13:53:02.980473 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.031839 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.043652 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.065942 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.099262 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.284508 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.298066 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.299710 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.372707 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.385988 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.471252 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.499138 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.592144 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.684298 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.714325 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.772322 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.811433 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.841343 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.861811 4848 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.869153 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dmbbx","openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.869241 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.877127 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.893176 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.893155914 podStartE2EDuration="20.893155914s" podCreationTimestamp="2025-12-04 13:52:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:53:03.892390315 +0000 UTC m=+287.834886893" watchObservedRunningTime="2025-12-04 13:53:03.893155914 +0000 UTC m=+287.835652452" Dec 04 13:53:03 crc kubenswrapper[4848]: I1204 13:53:03.967069 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.102467 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.161044 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.168095 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.232612 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.268101 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.304260 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.369416 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.382415 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.401619 4848 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.404442 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" path="/var/lib/kubelet/pods/eefea8cf-d6b0-45bf-8be8-1abf545d0e5c/volumes" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.479667 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.489311 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.503063 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.592460 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.605050 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.665898 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.794067 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.878468 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 13:53:04 crc kubenswrapper[4848]: I1204 13:53:04.888176 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.058364 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.070775 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.084987 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.211251 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.377511 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.432448 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.463234 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.520682 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.524216 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.543526 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.660055 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.698046 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.722184 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.759367 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.894095 4848 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.894384 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088" gracePeriod=5 Dec 04 13:53:05 crc kubenswrapper[4848]: I1204 13:53:05.955095 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.071681 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.113721 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.159330 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.199287 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.238426 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.362735 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.370121 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.377813 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.488537 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.677566 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.718883 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.776160 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.891601 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 13:53:06 crc kubenswrapper[4848]: I1204 13:53:06.982546 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.107437 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.136162 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.159314 4848 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.202873 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.206402 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.242061 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.344451 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.371542 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.372184 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.373257 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.376303 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.404438 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.437930 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.512813 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.515918 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.527932 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.539788 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.556258 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.706297 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.807755 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.823162 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.880879 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 13:53:07 crc kubenswrapper[4848]: I1204 13:53:07.892225 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.003307 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.044162 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.082238 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.122795 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.140811 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.224897 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.269258 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.314226 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.378781 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.509979 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.539169 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.545168 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.681488 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.761871 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.811512 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.923811 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5477954dc8-rmr8z"] Dec 04 13:53:08 crc kubenswrapper[4848]: E1204 13:53:08.924119 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" containerName="installer" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.924140 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" containerName="installer" Dec 04 13:53:08 crc kubenswrapper[4848]: E1204 13:53:08.924160 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.924173 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 13:53:08 crc kubenswrapper[4848]: E1204 13:53:08.924199 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" containerName="oauth-openshift" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.924214 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" containerName="oauth-openshift" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.924386 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.924414 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="eefea8cf-d6b0-45bf-8be8-1abf545d0e5c" containerName="oauth-openshift" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.924429 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a33ea42-e76c-4853-b7ad-68f96192d8f4" containerName="installer" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.925407 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.930208 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.931199 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.931246 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.931305 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.932301 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.935245 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.935354 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.935409 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.935973 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.936244 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.937817 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.943822 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.950546 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.952144 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.959681 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.967716 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 13:53:08 crc kubenswrapper[4848]: I1204 13:53:08.987502 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.007678 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.010446 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5477954dc8-rmr8z"] Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021115 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-error\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021182 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-session\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021218 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021255 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-login\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021306 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a56e6c00-eb57-4b3d-8692-0b188a643fc9-audit-dir\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021338 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021384 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021429 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021468 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021619 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021680 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qbgc\" (UniqueName: \"kubernetes.io/projected/a56e6c00-eb57-4b3d-8692-0b188a643fc9-kube-api-access-6qbgc\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021734 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021798 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.021847 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-audit-policies\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.037299 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.069472 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.122703 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.122755 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.122786 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.122843 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.122867 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qbgc\" (UniqueName: \"kubernetes.io/projected/a56e6c00-eb57-4b3d-8692-0b188a643fc9-kube-api-access-6qbgc\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.122890 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.122917 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.122941 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-audit-policies\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.123034 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-error\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.123061 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-session\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.123080 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.123104 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-login\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.123129 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a56e6c00-eb57-4b3d-8692-0b188a643fc9-audit-dir\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.123150 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.124384 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a56e6c00-eb57-4b3d-8692-0b188a643fc9-audit-dir\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.124981 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-audit-policies\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.128096 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.128782 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.129336 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.132850 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.132865 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-error\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.133192 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.134194 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-template-login\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.134962 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.136388 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.137765 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.138806 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a56e6c00-eb57-4b3d-8692-0b188a643fc9-v4-0-config-system-session\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.157843 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qbgc\" (UniqueName: \"kubernetes.io/projected/a56e6c00-eb57-4b3d-8692-0b188a643fc9-kube-api-access-6qbgc\") pod \"oauth-openshift-5477954dc8-rmr8z\" (UID: \"a56e6c00-eb57-4b3d-8692-0b188a643fc9\") " pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.169781 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.235486 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.250619 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.292801 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.330347 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.465607 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.480920 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.486200 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.624850 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.660446 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.698089 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.716703 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5477954dc8-rmr8z"] Dec 04 13:53:09 crc kubenswrapper[4848]: W1204 13:53:09.728542 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda56e6c00_eb57_4b3d_8692_0b188a643fc9.slice/crio-03c2520c777b55d5667f0e86d2759825feba4851554add6139bd32055bda62ee WatchSource:0}: Error finding container 03c2520c777b55d5667f0e86d2759825feba4851554add6139bd32055bda62ee: Status 404 returned error can't find the container with id 03c2520c777b55d5667f0e86d2759825feba4851554add6139bd32055bda62ee Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.776283 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 13:53:09 crc kubenswrapper[4848]: I1204 13:53:09.969079 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.057233 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.061846 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.079130 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.452048 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.537999 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.547426 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" event={"ID":"a56e6c00-eb57-4b3d-8692-0b188a643fc9","Type":"ContainerStarted","Data":"27b875b64986259bc532afc8bd97e87f3532a3101ad038da18e54eb403d364f2"} Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.547515 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" event={"ID":"a56e6c00-eb57-4b3d-8692-0b188a643fc9","Type":"ContainerStarted","Data":"03c2520c777b55d5667f0e86d2759825feba4851554add6139bd32055bda62ee"} Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.549518 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.575403 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.590356 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5477954dc8-rmr8z" podStartSLOduration=51.590335369 podStartE2EDuration="51.590335369s" podCreationTimestamp="2025-12-04 13:52:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:53:10.586140964 +0000 UTC m=+294.528637512" watchObservedRunningTime="2025-12-04 13:53:10.590335369 +0000 UTC m=+294.532831907" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.825178 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.841587 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.892634 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 13:53:10 crc kubenswrapper[4848]: I1204 13:53:10.977381 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.043281 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.080671 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.221623 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.256126 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.463781 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.505172 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.505270 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.528250 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.554590 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.554662 4848 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088" exitCode=137 Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.554789 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.554891 4848 scope.go:117] "RemoveContainer" containerID="689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572094 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572186 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572236 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572310 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572426 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572489 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572484 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572553 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572688 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.572938 4848 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.573046 4848 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.573066 4848 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.573085 4848 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.583528 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.585810 4848 scope.go:117] "RemoveContainer" containerID="689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088" Dec 04 13:53:11 crc kubenswrapper[4848]: E1204 13:53:11.586519 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088\": container with ID starting with 689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088 not found: ID does not exist" containerID="689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.586570 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088"} err="failed to get container status \"689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088\": rpc error: code = NotFound desc = could not find container \"689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088\": container with ID starting with 689d7f16bf58c9ac5295b1e44cbcf8d7dc62c501dfcf7873737f1d52f2d6e088 not found: ID does not exist" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.675662 4848 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.879817 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.917516 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.921756 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 13:53:11 crc kubenswrapper[4848]: I1204 13:53:11.930832 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 13:53:12 crc kubenswrapper[4848]: I1204 13:53:12.407106 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 04 13:53:12 crc kubenswrapper[4848]: I1204 13:53:12.491527 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 13:53:16 crc kubenswrapper[4848]: I1204 13:53:16.217625 4848 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.330460 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sd8h2"] Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.331186 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" podUID="aed44f8e-efc1-4a6d-be21-e97dee631132" containerName="controller-manager" containerID="cri-o://17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14" gracePeriod=30 Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.387563 4848 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-sd8h2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.387824 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" podUID="aed44f8e-efc1-4a6d-be21-e97dee631132" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.437605 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8"] Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.437843 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" podUID="82ae395f-c4ee-43de-a1d6-0349c6dc640b" containerName="route-controller-manager" containerID="cri-o://d1179f54587ec46800449bc4211a16305ff611d9ecc3eb4c092c941a3ecc21c9" gracePeriod=30 Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.663324 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.765460 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed44f8e-efc1-4a6d-be21-e97dee631132-serving-cert\") pod \"aed44f8e-efc1-4a6d-be21-e97dee631132\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.765548 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-config\") pod \"aed44f8e-efc1-4a6d-be21-e97dee631132\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.765575 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-proxy-ca-bundles\") pod \"aed44f8e-efc1-4a6d-be21-e97dee631132\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.765669 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-client-ca\") pod \"aed44f8e-efc1-4a6d-be21-e97dee631132\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.765718 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwp9h\" (UniqueName: \"kubernetes.io/projected/aed44f8e-efc1-4a6d-be21-e97dee631132-kube-api-access-gwp9h\") pod \"aed44f8e-efc1-4a6d-be21-e97dee631132\" (UID: \"aed44f8e-efc1-4a6d-be21-e97dee631132\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.766807 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-config" (OuterVolumeSpecName: "config") pod "aed44f8e-efc1-4a6d-be21-e97dee631132" (UID: "aed44f8e-efc1-4a6d-be21-e97dee631132"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.767267 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "aed44f8e-efc1-4a6d-be21-e97dee631132" (UID: "aed44f8e-efc1-4a6d-be21-e97dee631132"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.767396 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-client-ca" (OuterVolumeSpecName: "client-ca") pod "aed44f8e-efc1-4a6d-be21-e97dee631132" (UID: "aed44f8e-efc1-4a6d-be21-e97dee631132"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.776944 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed44f8e-efc1-4a6d-be21-e97dee631132-kube-api-access-gwp9h" (OuterVolumeSpecName: "kube-api-access-gwp9h") pod "aed44f8e-efc1-4a6d-be21-e97dee631132" (UID: "aed44f8e-efc1-4a6d-be21-e97dee631132"). InnerVolumeSpecName "kube-api-access-gwp9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.777870 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed44f8e-efc1-4a6d-be21-e97dee631132-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "aed44f8e-efc1-4a6d-be21-e97dee631132" (UID: "aed44f8e-efc1-4a6d-be21-e97dee631132"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.781753 4848 generic.go:334] "Generic (PLEG): container finished" podID="aed44f8e-efc1-4a6d-be21-e97dee631132" containerID="17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14" exitCode=0 Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.781807 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.781831 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" event={"ID":"aed44f8e-efc1-4a6d-be21-e97dee631132","Type":"ContainerDied","Data":"17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14"} Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.781862 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sd8h2" event={"ID":"aed44f8e-efc1-4a6d-be21-e97dee631132","Type":"ContainerDied","Data":"04368de60201de9764f1a502c5a4b2a51ab39c234239aba9842779c036f49157"} Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.781880 4848 scope.go:117] "RemoveContainer" containerID="17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.785716 4848 generic.go:334] "Generic (PLEG): container finished" podID="82ae395f-c4ee-43de-a1d6-0349c6dc640b" containerID="d1179f54587ec46800449bc4211a16305ff611d9ecc3eb4c092c941a3ecc21c9" exitCode=0 Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.785756 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" event={"ID":"82ae395f-c4ee-43de-a1d6-0349c6dc640b","Type":"ContainerDied","Data":"d1179f54587ec46800449bc4211a16305ff611d9ecc3eb4c092c941a3ecc21c9"} Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.808186 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sd8h2"] Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.808749 4848 scope.go:117] "RemoveContainer" containerID="17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14" Dec 04 13:53:45 crc kubenswrapper[4848]: E1204 13:53:45.809301 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14\": container with ID starting with 17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14 not found: ID does not exist" containerID="17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.809334 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14"} err="failed to get container status \"17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14\": rpc error: code = NotFound desc = could not find container \"17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14\": container with ID starting with 17f0b0a178032f044a5bf3494a9bf22c8b45579693c53ef7ccdccc43dba33a14 not found: ID does not exist" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.828036 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sd8h2"] Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.844102 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.868136 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwp9h\" (UniqueName: \"kubernetes.io/projected/aed44f8e-efc1-4a6d-be21-e97dee631132-kube-api-access-gwp9h\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.868171 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aed44f8e-efc1-4a6d-be21-e97dee631132-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.868201 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.868216 4848 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.868227 4848 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aed44f8e-efc1-4a6d-be21-e97dee631132-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.969449 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ae395f-c4ee-43de-a1d6-0349c6dc640b-serving-cert\") pod \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.969637 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp6sd\" (UniqueName: \"kubernetes.io/projected/82ae395f-c4ee-43de-a1d6-0349c6dc640b-kube-api-access-zp6sd\") pod \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.969675 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-config\") pod \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.969700 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-client-ca\") pod \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\" (UID: \"82ae395f-c4ee-43de-a1d6-0349c6dc640b\") " Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.970235 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-client-ca" (OuterVolumeSpecName: "client-ca") pod "82ae395f-c4ee-43de-a1d6-0349c6dc640b" (UID: "82ae395f-c4ee-43de-a1d6-0349c6dc640b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.970332 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-config" (OuterVolumeSpecName: "config") pod "82ae395f-c4ee-43de-a1d6-0349c6dc640b" (UID: "82ae395f-c4ee-43de-a1d6-0349c6dc640b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.972577 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ae395f-c4ee-43de-a1d6-0349c6dc640b-kube-api-access-zp6sd" (OuterVolumeSpecName: "kube-api-access-zp6sd") pod "82ae395f-c4ee-43de-a1d6-0349c6dc640b" (UID: "82ae395f-c4ee-43de-a1d6-0349c6dc640b"). InnerVolumeSpecName "kube-api-access-zp6sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:53:45 crc kubenswrapper[4848]: I1204 13:53:45.972803 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ae395f-c4ee-43de-a1d6-0349c6dc640b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "82ae395f-c4ee-43de-a1d6-0349c6dc640b" (UID: "82ae395f-c4ee-43de-a1d6-0349c6dc640b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.071112 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp6sd\" (UniqueName: \"kubernetes.io/projected/82ae395f-c4ee-43de-a1d6-0349c6dc640b-kube-api-access-zp6sd\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.071143 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.071152 4848 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82ae395f-c4ee-43de-a1d6-0349c6dc640b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.071160 4848 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ae395f-c4ee-43de-a1d6-0349c6dc640b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.406166 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed44f8e-efc1-4a6d-be21-e97dee631132" path="/var/lib/kubelet/pods/aed44f8e-efc1-4a6d-be21-e97dee631132/volumes" Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.800923 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" event={"ID":"82ae395f-c4ee-43de-a1d6-0349c6dc640b","Type":"ContainerDied","Data":"9bccd07947f67859f560f3f6d4906ce44127d6b6b10e29168ce6c67478af3091"} Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.801487 4848 scope.go:117] "RemoveContainer" containerID="d1179f54587ec46800449bc4211a16305ff611d9ecc3eb4c092c941a3ecc21c9" Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.801010 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8" Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.826157 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8"] Dec 04 13:53:46 crc kubenswrapper[4848]: I1204 13:53:46.830745 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lwll8"] Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.265726 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bfd8b4665-szsx5"] Dec 04 13:53:47 crc kubenswrapper[4848]: E1204 13:53:47.266531 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed44f8e-efc1-4a6d-be21-e97dee631132" containerName="controller-manager" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.268395 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed44f8e-efc1-4a6d-be21-e97dee631132" containerName="controller-manager" Dec 04 13:53:47 crc kubenswrapper[4848]: E1204 13:53:47.268601 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ae395f-c4ee-43de-a1d6-0349c6dc640b" containerName="route-controller-manager" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.268735 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ae395f-c4ee-43de-a1d6-0349c6dc640b" containerName="route-controller-manager" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.270133 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ae395f-c4ee-43de-a1d6-0349c6dc640b" containerName="route-controller-manager" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.270240 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed44f8e-efc1-4a6d-be21-e97dee631132" containerName="controller-manager" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.270883 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.274288 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw"] Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.274470 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.275868 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.276810 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.276809 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.276988 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.277094 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.279166 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.280801 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.280813 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.281031 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.280973 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.281726 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bfd8b4665-szsx5"] Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.290582 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.292437 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.297770 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.302730 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw"] Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.389914 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661dabe6-bd9e-4468-89c0-cdb922a64d02-config\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.389991 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90551a0a-b873-420d-bbac-995309b42c8b-serving-cert\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.390022 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-config\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.390051 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcw87\" (UniqueName: \"kubernetes.io/projected/661dabe6-bd9e-4468-89c0-cdb922a64d02-kube-api-access-qcw87\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.390074 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-client-ca\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.390203 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/661dabe6-bd9e-4468-89c0-cdb922a64d02-client-ca\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.390286 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8kqn\" (UniqueName: \"kubernetes.io/projected/90551a0a-b873-420d-bbac-995309b42c8b-kube-api-access-l8kqn\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.390345 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/661dabe6-bd9e-4468-89c0-cdb922a64d02-serving-cert\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.390459 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-proxy-ca-bundles\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491504 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-client-ca\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491569 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/661dabe6-bd9e-4468-89c0-cdb922a64d02-client-ca\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491614 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8kqn\" (UniqueName: \"kubernetes.io/projected/90551a0a-b873-420d-bbac-995309b42c8b-kube-api-access-l8kqn\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491652 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/661dabe6-bd9e-4468-89c0-cdb922a64d02-serving-cert\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491684 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-proxy-ca-bundles\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491721 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661dabe6-bd9e-4468-89c0-cdb922a64d02-config\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491753 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90551a0a-b873-420d-bbac-995309b42c8b-serving-cert\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491813 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-config\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.491865 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcw87\" (UniqueName: \"kubernetes.io/projected/661dabe6-bd9e-4468-89c0-cdb922a64d02-kube-api-access-qcw87\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.493026 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-client-ca\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.493281 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/661dabe6-bd9e-4468-89c0-cdb922a64d02-client-ca\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.493581 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-proxy-ca-bundles\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.494445 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661dabe6-bd9e-4468-89c0-cdb922a64d02-config\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.495683 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90551a0a-b873-420d-bbac-995309b42c8b-config\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.500630 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/661dabe6-bd9e-4468-89c0-cdb922a64d02-serving-cert\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.506778 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90551a0a-b873-420d-bbac-995309b42c8b-serving-cert\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.514908 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcw87\" (UniqueName: \"kubernetes.io/projected/661dabe6-bd9e-4468-89c0-cdb922a64d02-kube-api-access-qcw87\") pod \"route-controller-manager-5668ccbb96-4khmw\" (UID: \"661dabe6-bd9e-4468-89c0-cdb922a64d02\") " pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.515054 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8kqn\" (UniqueName: \"kubernetes.io/projected/90551a0a-b873-420d-bbac-995309b42c8b-kube-api-access-l8kqn\") pod \"controller-manager-6bfd8b4665-szsx5\" (UID: \"90551a0a-b873-420d-bbac-995309b42c8b\") " pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.593237 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.609515 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:47 crc kubenswrapper[4848]: I1204 13:53:47.826254 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw"] Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.015774 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bfd8b4665-szsx5"] Dec 04 13:53:48 crc kubenswrapper[4848]: W1204 13:53:48.017595 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90551a0a_b873_420d_bbac_995309b42c8b.slice/crio-99bf7fe3e00033ed2f546dc2da87f1fddfee6960b3252a4db909218a33b0fdb2 WatchSource:0}: Error finding container 99bf7fe3e00033ed2f546dc2da87f1fddfee6960b3252a4db909218a33b0fdb2: Status 404 returned error can't find the container with id 99bf7fe3e00033ed2f546dc2da87f1fddfee6960b3252a4db909218a33b0fdb2 Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.408777 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ae395f-c4ee-43de-a1d6-0349c6dc640b" path="/var/lib/kubelet/pods/82ae395f-c4ee-43de-a1d6-0349c6dc640b/volumes" Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.815230 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" event={"ID":"90551a0a-b873-420d-bbac-995309b42c8b","Type":"ContainerStarted","Data":"c1b6002edb4da9044c1d2837c0c503b78ab9209bb1275e752434ad6829f19076"} Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.815279 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" event={"ID":"90551a0a-b873-420d-bbac-995309b42c8b","Type":"ContainerStarted","Data":"99bf7fe3e00033ed2f546dc2da87f1fddfee6960b3252a4db909218a33b0fdb2"} Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.815463 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.816626 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" event={"ID":"661dabe6-bd9e-4468-89c0-cdb922a64d02","Type":"ContainerStarted","Data":"6f4a056e32050757562fb68f5c558673d2f287338f89d7671cb2bb63597a8643"} Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.816690 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" event={"ID":"661dabe6-bd9e-4468-89c0-cdb922a64d02","Type":"ContainerStarted","Data":"50cbdd88e6db69dd7e5b509f2b6fe3ae693bedb8a5bd92baa20bec75ad36ea2b"} Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.819356 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.824324 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.825104 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.839778 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bfd8b4665-szsx5" podStartSLOduration=3.839762103 podStartE2EDuration="3.839762103s" podCreationTimestamp="2025-12-04 13:53:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:53:48.835481207 +0000 UTC m=+332.777977745" watchObservedRunningTime="2025-12-04 13:53:48.839762103 +0000 UTC m=+332.782258631" Dec 04 13:53:48 crc kubenswrapper[4848]: I1204 13:53:48.870410 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5668ccbb96-4khmw" podStartSLOduration=3.8703745019999998 podStartE2EDuration="3.870374502s" podCreationTimestamp="2025-12-04 13:53:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:53:48.863758961 +0000 UTC m=+332.806255529" watchObservedRunningTime="2025-12-04 13:53:48.870374502 +0000 UTC m=+332.812871040" Dec 04 13:54:14 crc kubenswrapper[4848]: I1204 13:54:14.314524 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:54:14 crc kubenswrapper[4848]: I1204 13:54:14.315191 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.978353 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8x6gs"] Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.979464 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.994843 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-trusted-ca\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.994911 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-registry-certificates\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.994941 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.995008 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.995029 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-registry-tls\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.995051 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.995081 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-bound-sa-token\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:37 crc kubenswrapper[4848]: I1204 13:54:37.995148 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltlt8\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-kube-api-access-ltlt8\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.003039 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8x6gs"] Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.035253 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.107855 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-trusted-ca\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.108106 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-registry-certificates\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.108194 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.108241 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-registry-tls\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.108283 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.108325 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-bound-sa-token\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.108379 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltlt8\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-kube-api-access-ltlt8\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.109454 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.109464 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-trusted-ca\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.109620 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-registry-certificates\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.114344 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.115851 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-registry-tls\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.131761 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltlt8\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-kube-api-access-ltlt8\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.134672 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b9f3490d-5fab-4e7b-9fdc-42e76356ead8-bound-sa-token\") pod \"image-registry-66df7c8f76-8x6gs\" (UID: \"b9f3490d-5fab-4e7b-9fdc-42e76356ead8\") " pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.293693 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:38 crc kubenswrapper[4848]: I1204 13:54:38.702511 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8x6gs"] Dec 04 13:54:39 crc kubenswrapper[4848]: I1204 13:54:39.112466 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" event={"ID":"b9f3490d-5fab-4e7b-9fdc-42e76356ead8","Type":"ContainerStarted","Data":"794bf1bb9b07055d8774bbbdb16900da9053df4e76ffbfd2972676ea3864a01e"} Dec 04 13:54:39 crc kubenswrapper[4848]: I1204 13:54:39.112752 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" event={"ID":"b9f3490d-5fab-4e7b-9fdc-42e76356ead8","Type":"ContainerStarted","Data":"4995218b17cee3047c4a177940c4a2a02af78b94856ae7096475a0e7502e4500"} Dec 04 13:54:39 crc kubenswrapper[4848]: I1204 13:54:39.112767 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:39 crc kubenswrapper[4848]: I1204 13:54:39.129843 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" podStartSLOduration=2.129827613 podStartE2EDuration="2.129827613s" podCreationTimestamp="2025-12-04 13:54:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:54:39.128786748 +0000 UTC m=+383.071283296" watchObservedRunningTime="2025-12-04 13:54:39.129827613 +0000 UTC m=+383.072324141" Dec 04 13:54:44 crc kubenswrapper[4848]: I1204 13:54:44.313917 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:54:44 crc kubenswrapper[4848]: I1204 13:54:44.314425 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.802726 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjljj"] Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.803791 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sjljj" podUID="e75ab632-da17-4354-951c-476ad2e46835" containerName="registry-server" containerID="cri-o://0a090d37c3da4dbf63f6d6e5ade249782fd8254f8e48cd6c92e463c134cc8324" gracePeriod=30 Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.810333 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nr5qh"] Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.810878 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nr5qh" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" containerName="registry-server" containerID="cri-o://3059dc72ca92d3456e72586b9cf129c03bf7347e00676d8c1483eb860d8eb816" gracePeriod=30 Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.820979 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vv5h2"] Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.821206 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" podUID="74527b90-33e5-4580-bee4-e71afc8a89f7" containerName="marketplace-operator" containerID="cri-o://f6b502d805ba0c2e60972126a16c489a7ba21a7c968d30447eee6fea5a87b08d" gracePeriod=30 Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.829234 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrx58"] Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.829599 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hrx58" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" containerName="registry-server" containerID="cri-o://39abe968350d51fd71413c221f1bda289f39a61599b759f26a4669796633bc1e" gracePeriod=30 Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.836740 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x6h7q"] Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.837054 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x6h7q" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="registry-server" containerID="cri-o://7109be40b4f575f22c41ac3d20b08d6aa828a36c19d78c63fd064d8000571b37" gracePeriod=30 Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.849489 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nq9tw"] Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.850618 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.868866 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nq9tw"] Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.955130 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cf33e30f-cf18-4cdd-8b35-84f835998ffd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.955191 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cf33e30f-cf18-4cdd-8b35-84f835998ffd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:54 crc kubenswrapper[4848]: I1204 13:54:54.955244 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js4pm\" (UniqueName: \"kubernetes.io/projected/cf33e30f-cf18-4cdd-8b35-84f835998ffd-kube-api-access-js4pm\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.056446 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cf33e30f-cf18-4cdd-8b35-84f835998ffd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.056500 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cf33e30f-cf18-4cdd-8b35-84f835998ffd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.056534 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js4pm\" (UniqueName: \"kubernetes.io/projected/cf33e30f-cf18-4cdd-8b35-84f835998ffd-kube-api-access-js4pm\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.058680 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cf33e30f-cf18-4cdd-8b35-84f835998ffd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.062707 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cf33e30f-cf18-4cdd-8b35-84f835998ffd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.072871 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js4pm\" (UniqueName: \"kubernetes.io/projected/cf33e30f-cf18-4cdd-8b35-84f835998ffd-kube-api-access-js4pm\") pod \"marketplace-operator-79b997595-nq9tw\" (UID: \"cf33e30f-cf18-4cdd-8b35-84f835998ffd\") " pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.220016 4848 generic.go:334] "Generic (PLEG): container finished" podID="e75ab632-da17-4354-951c-476ad2e46835" containerID="0a090d37c3da4dbf63f6d6e5ade249782fd8254f8e48cd6c92e463c134cc8324" exitCode=0 Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.220089 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjljj" event={"ID":"e75ab632-da17-4354-951c-476ad2e46835","Type":"ContainerDied","Data":"0a090d37c3da4dbf63f6d6e5ade249782fd8254f8e48cd6c92e463c134cc8324"} Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.220117 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjljj" event={"ID":"e75ab632-da17-4354-951c-476ad2e46835","Type":"ContainerDied","Data":"215f327d5a8796c31cc9dc05de5f1ed6087d8a66b39ffe77470f075e80b3b2d4"} Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.220130 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="215f327d5a8796c31cc9dc05de5f1ed6087d8a66b39ffe77470f075e80b3b2d4" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.223268 4848 generic.go:334] "Generic (PLEG): container finished" podID="8eedf821-9e1a-4513-896f-372df7171ef2" containerID="39abe968350d51fd71413c221f1bda289f39a61599b759f26a4669796633bc1e" exitCode=0 Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.223332 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrx58" event={"ID":"8eedf821-9e1a-4513-896f-372df7171ef2","Type":"ContainerDied","Data":"39abe968350d51fd71413c221f1bda289f39a61599b759f26a4669796633bc1e"} Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.226105 4848 generic.go:334] "Generic (PLEG): container finished" podID="15f8738d-4240-42d0-a819-cccff5aaa821" containerID="7109be40b4f575f22c41ac3d20b08d6aa828a36c19d78c63fd064d8000571b37" exitCode=0 Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.226156 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6h7q" event={"ID":"15f8738d-4240-42d0-a819-cccff5aaa821","Type":"ContainerDied","Data":"7109be40b4f575f22c41ac3d20b08d6aa828a36c19d78c63fd064d8000571b37"} Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.229886 4848 generic.go:334] "Generic (PLEG): container finished" podID="74527b90-33e5-4580-bee4-e71afc8a89f7" containerID="f6b502d805ba0c2e60972126a16c489a7ba21a7c968d30447eee6fea5a87b08d" exitCode=0 Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.229979 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" event={"ID":"74527b90-33e5-4580-bee4-e71afc8a89f7","Type":"ContainerDied","Data":"f6b502d805ba0c2e60972126a16c489a7ba21a7c968d30447eee6fea5a87b08d"} Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.230018 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.237454 4848 generic.go:334] "Generic (PLEG): container finished" podID="f3d120a8-d58e-45b9-be60-04be599332b3" containerID="3059dc72ca92d3456e72586b9cf129c03bf7347e00676d8c1483eb860d8eb816" exitCode=0 Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.237536 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5qh" event={"ID":"f3d120a8-d58e-45b9-be60-04be599332b3","Type":"ContainerDied","Data":"3059dc72ca92d3456e72586b9cf129c03bf7347e00676d8c1483eb860d8eb816"} Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.242776 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.288364 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.291839 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.310902 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.315491 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.364710 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-catalog-content\") pod \"e75ab632-da17-4354-951c-476ad2e46835\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.364780 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgbzl\" (UniqueName: \"kubernetes.io/projected/e75ab632-da17-4354-951c-476ad2e46835-kube-api-access-lgbzl\") pod \"e75ab632-da17-4354-951c-476ad2e46835\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.364806 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-utilities\") pod \"e75ab632-da17-4354-951c-476ad2e46835\" (UID: \"e75ab632-da17-4354-951c-476ad2e46835\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.370901 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-utilities" (OuterVolumeSpecName: "utilities") pod "e75ab632-da17-4354-951c-476ad2e46835" (UID: "e75ab632-da17-4354-951c-476ad2e46835"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.385482 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75ab632-da17-4354-951c-476ad2e46835-kube-api-access-lgbzl" (OuterVolumeSpecName: "kube-api-access-lgbzl") pod "e75ab632-da17-4354-951c-476ad2e46835" (UID: "e75ab632-da17-4354-951c-476ad2e46835"). InnerVolumeSpecName "kube-api-access-lgbzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.426643 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e75ab632-da17-4354-951c-476ad2e46835" (UID: "e75ab632-da17-4354-951c-476ad2e46835"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466061 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87tcj\" (UniqueName: \"kubernetes.io/projected/74527b90-33e5-4580-bee4-e71afc8a89f7-kube-api-access-87tcj\") pod \"74527b90-33e5-4580-bee4-e71afc8a89f7\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466122 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-utilities\") pod \"8eedf821-9e1a-4513-896f-372df7171ef2\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466146 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89jwl\" (UniqueName: \"kubernetes.io/projected/15f8738d-4240-42d0-a819-cccff5aaa821-kube-api-access-89jwl\") pod \"15f8738d-4240-42d0-a819-cccff5aaa821\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466164 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgk26\" (UniqueName: \"kubernetes.io/projected/f3d120a8-d58e-45b9-be60-04be599332b3-kube-api-access-fgk26\") pod \"f3d120a8-d58e-45b9-be60-04be599332b3\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466225 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-trusted-ca\") pod \"74527b90-33e5-4580-bee4-e71afc8a89f7\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466266 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-catalog-content\") pod \"15f8738d-4240-42d0-a819-cccff5aaa821\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466294 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-utilities\") pod \"15f8738d-4240-42d0-a819-cccff5aaa821\" (UID: \"15f8738d-4240-42d0-a819-cccff5aaa821\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466346 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-utilities\") pod \"f3d120a8-d58e-45b9-be60-04be599332b3\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466364 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-catalog-content\") pod \"8eedf821-9e1a-4513-896f-372df7171ef2\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466392 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4zvc\" (UniqueName: \"kubernetes.io/projected/8eedf821-9e1a-4513-896f-372df7171ef2-kube-api-access-k4zvc\") pod \"8eedf821-9e1a-4513-896f-372df7171ef2\" (UID: \"8eedf821-9e1a-4513-896f-372df7171ef2\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466423 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-operator-metrics\") pod \"74527b90-33e5-4580-bee4-e71afc8a89f7\" (UID: \"74527b90-33e5-4580-bee4-e71afc8a89f7\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466441 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-catalog-content\") pod \"f3d120a8-d58e-45b9-be60-04be599332b3\" (UID: \"f3d120a8-d58e-45b9-be60-04be599332b3\") " Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466689 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466702 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgbzl\" (UniqueName: \"kubernetes.io/projected/e75ab632-da17-4354-951c-476ad2e46835-kube-api-access-lgbzl\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.466712 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e75ab632-da17-4354-951c-476ad2e46835-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.467192 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-utilities" (OuterVolumeSpecName: "utilities") pod "8eedf821-9e1a-4513-896f-372df7171ef2" (UID: "8eedf821-9e1a-4513-896f-372df7171ef2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.467835 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-utilities" (OuterVolumeSpecName: "utilities") pod "15f8738d-4240-42d0-a819-cccff5aaa821" (UID: "15f8738d-4240-42d0-a819-cccff5aaa821"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.467865 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "74527b90-33e5-4580-bee4-e71afc8a89f7" (UID: "74527b90-33e5-4580-bee4-e71afc8a89f7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.469062 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-utilities" (OuterVolumeSpecName: "utilities") pod "f3d120a8-d58e-45b9-be60-04be599332b3" (UID: "f3d120a8-d58e-45b9-be60-04be599332b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.469913 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15f8738d-4240-42d0-a819-cccff5aaa821-kube-api-access-89jwl" (OuterVolumeSpecName: "kube-api-access-89jwl") pod "15f8738d-4240-42d0-a819-cccff5aaa821" (UID: "15f8738d-4240-42d0-a819-cccff5aaa821"). InnerVolumeSpecName "kube-api-access-89jwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.469902 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "74527b90-33e5-4580-bee4-e71afc8a89f7" (UID: "74527b90-33e5-4580-bee4-e71afc8a89f7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.470079 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74527b90-33e5-4580-bee4-e71afc8a89f7-kube-api-access-87tcj" (OuterVolumeSpecName: "kube-api-access-87tcj") pod "74527b90-33e5-4580-bee4-e71afc8a89f7" (UID: "74527b90-33e5-4580-bee4-e71afc8a89f7"). InnerVolumeSpecName "kube-api-access-87tcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.470603 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3d120a8-d58e-45b9-be60-04be599332b3-kube-api-access-fgk26" (OuterVolumeSpecName: "kube-api-access-fgk26") pod "f3d120a8-d58e-45b9-be60-04be599332b3" (UID: "f3d120a8-d58e-45b9-be60-04be599332b3"). InnerVolumeSpecName "kube-api-access-fgk26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.471223 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eedf821-9e1a-4513-896f-372df7171ef2-kube-api-access-k4zvc" (OuterVolumeSpecName: "kube-api-access-k4zvc") pod "8eedf821-9e1a-4513-896f-372df7171ef2" (UID: "8eedf821-9e1a-4513-896f-372df7171ef2"). InnerVolumeSpecName "kube-api-access-k4zvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.486408 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8eedf821-9e1a-4513-896f-372df7171ef2" (UID: "8eedf821-9e1a-4513-896f-372df7171ef2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.522101 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3d120a8-d58e-45b9-be60-04be599332b3" (UID: "f3d120a8-d58e-45b9-be60-04be599332b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568522 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4zvc\" (UniqueName: \"kubernetes.io/projected/8eedf821-9e1a-4513-896f-372df7171ef2-kube-api-access-k4zvc\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568564 4848 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568594 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568657 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87tcj\" (UniqueName: \"kubernetes.io/projected/74527b90-33e5-4580-bee4-e71afc8a89f7-kube-api-access-87tcj\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568676 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568700 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89jwl\" (UniqueName: \"kubernetes.io/projected/15f8738d-4240-42d0-a819-cccff5aaa821-kube-api-access-89jwl\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568723 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgk26\" (UniqueName: \"kubernetes.io/projected/f3d120a8-d58e-45b9-be60-04be599332b3-kube-api-access-fgk26\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568740 4848 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74527b90-33e5-4580-bee4-e71afc8a89f7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568755 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568770 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eedf821-9e1a-4513-896f-372df7171ef2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.568786 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3d120a8-d58e-45b9-be60-04be599332b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.577833 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15f8738d-4240-42d0-a819-cccff5aaa821" (UID: "15f8738d-4240-42d0-a819-cccff5aaa821"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.670025 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f8738d-4240-42d0-a819-cccff5aaa821-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:54:55 crc kubenswrapper[4848]: I1204 13:54:55.682617 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nq9tw"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.244527 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x6h7q" event={"ID":"15f8738d-4240-42d0-a819-cccff5aaa821","Type":"ContainerDied","Data":"dd10f453727a3f8bab92f925578ca42bb79d22e440d3e3fdd271deda31cf114d"} Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.244550 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x6h7q" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.244591 4848 scope.go:117] "RemoveContainer" containerID="7109be40b4f575f22c41ac3d20b08d6aa828a36c19d78c63fd064d8000571b37" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.246346 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" event={"ID":"74527b90-33e5-4580-bee4-e71afc8a89f7","Type":"ContainerDied","Data":"1a5d238e545c34525ddacd36fa2650e7641fb7c5b3a9f3c99bae5c70be1751b8"} Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.246422 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vv5h2" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.248427 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nr5qh" event={"ID":"f3d120a8-d58e-45b9-be60-04be599332b3","Type":"ContainerDied","Data":"e7c3d6c69e28d6584798e9212d42a275bd0cc7c38744a4434df2924e743ef96f"} Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.249110 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nr5qh" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.253784 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrx58" event={"ID":"8eedf821-9e1a-4513-896f-372df7171ef2","Type":"ContainerDied","Data":"9914984ebb7744afd30a678f6134757d9147b5adc576b03e36a767090d4975f9"} Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.253878 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrx58" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.255843 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjljj" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.256349 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" event={"ID":"cf33e30f-cf18-4cdd-8b35-84f835998ffd","Type":"ContainerStarted","Data":"b578e924881eaca7cb56f43ac7db77d3899b880170f14d070c9cb0c925ef6343"} Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.256417 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" event={"ID":"cf33e30f-cf18-4cdd-8b35-84f835998ffd","Type":"ContainerStarted","Data":"ecc7cb88aa1119f3ed1bce457b1e08294c1e7077ef2108b87f586151d33eb2e2"} Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.256440 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.261396 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.267872 4848 scope.go:117] "RemoveContainer" containerID="9d178022a3b94d4aabe32a720b73dfedd5c341214b835068b82a7800237add6d" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.287084 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-nq9tw" podStartSLOduration=2.287065806 podStartE2EDuration="2.287065806s" podCreationTimestamp="2025-12-04 13:54:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:54:56.281843589 +0000 UTC m=+400.224340127" watchObservedRunningTime="2025-12-04 13:54:56.287065806 +0000 UTC m=+400.229562334" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.319085 4848 scope.go:117] "RemoveContainer" containerID="fa918e58186f5974a99e08cd4471f2311711c21a8e366c82b241acb4d6b48537" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.324523 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nr5qh"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.343317 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nr5qh"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.351594 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x6h7q"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.356114 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x6h7q"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.358234 4848 scope.go:117] "RemoveContainer" containerID="f6b502d805ba0c2e60972126a16c489a7ba21a7c968d30447eee6fea5a87b08d" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.371534 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjljj"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.376308 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sjljj"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.379168 4848 scope.go:117] "RemoveContainer" containerID="3059dc72ca92d3456e72586b9cf129c03bf7347e00676d8c1483eb860d8eb816" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.383191 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrx58"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.387615 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrx58"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.390665 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vv5h2"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.392701 4848 scope.go:117] "RemoveContainer" containerID="f23ee88bcfb4887ef5fbbe679b4811704df7344ed0f7ddd68b391a057bef8487" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.401814 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" path="/var/lib/kubelet/pods/15f8738d-4240-42d0-a819-cccff5aaa821/volumes" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.402533 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" path="/var/lib/kubelet/pods/8eedf821-9e1a-4513-896f-372df7171ef2/volumes" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.403266 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e75ab632-da17-4354-951c-476ad2e46835" path="/var/lib/kubelet/pods/e75ab632-da17-4354-951c-476ad2e46835/volumes" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.404745 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" path="/var/lib/kubelet/pods/f3d120a8-d58e-45b9-be60-04be599332b3/volumes" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.405356 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vv5h2"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.405359 4848 scope.go:117] "RemoveContainer" containerID="760ee0d53304a79ced9adbeed9c64f4c5f56137c58551b302a474e0402afed15" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.418656 4848 scope.go:117] "RemoveContainer" containerID="39abe968350d51fd71413c221f1bda289f39a61599b759f26a4669796633bc1e" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.433672 4848 scope.go:117] "RemoveContainer" containerID="3aef13f92912d30feed58b4eca566026cbb430ea49de87ad4bf8fc98f384a926" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.443836 4848 scope.go:117] "RemoveContainer" containerID="33c698f047dbcb19bd2cec40862123a70ff599ccb99d9490a551bedc163f26b4" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.805726 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rkwjn"] Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.805916 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e75ab632-da17-4354-951c-476ad2e46835" containerName="extract-utilities" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.805927 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="e75ab632-da17-4354-951c-476ad2e46835" containerName="extract-utilities" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.805934 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" containerName="extract-content" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.805940 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" containerName="extract-content" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.805967 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e75ab632-da17-4354-951c-476ad2e46835" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.805973 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="e75ab632-da17-4354-951c-476ad2e46835" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.805984 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" containerName="extract-content" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.805991 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" containerName="extract-content" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806004 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="extract-utilities" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806011 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="extract-utilities" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806020 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="extract-content" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806026 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="extract-content" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806037 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806042 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806050 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" containerName="extract-utilities" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806055 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" containerName="extract-utilities" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806064 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e75ab632-da17-4354-951c-476ad2e46835" containerName="extract-content" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806069 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="e75ab632-da17-4354-951c-476ad2e46835" containerName="extract-content" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806076 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74527b90-33e5-4580-bee4-e71afc8a89f7" containerName="marketplace-operator" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806102 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="74527b90-33e5-4580-bee4-e71afc8a89f7" containerName="marketplace-operator" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806110 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806115 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806123 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806129 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: E1204 13:54:56.806138 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" containerName="extract-utilities" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806143 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" containerName="extract-utilities" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806259 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f8738d-4240-42d0-a819-cccff5aaa821" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806272 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="74527b90-33e5-4580-bee4-e71afc8a89f7" containerName="marketplace-operator" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806280 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="e75ab632-da17-4354-951c-476ad2e46835" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806288 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eedf821-9e1a-4513-896f-372df7171ef2" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.806294 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3d120a8-d58e-45b9-be60-04be599332b3" containerName="registry-server" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.807070 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.817936 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkwjn"] Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.818471 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.984448 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-utilities\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.984699 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-catalog-content\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:56 crc kubenswrapper[4848]: I1204 13:54:56.984784 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9psb\" (UniqueName: \"kubernetes.io/projected/b8658e99-6b59-4722-90b2-7a18294722df-kube-api-access-x9psb\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.086730 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9psb\" (UniqueName: \"kubernetes.io/projected/b8658e99-6b59-4722-90b2-7a18294722df-kube-api-access-x9psb\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.086822 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-utilities\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.086870 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-catalog-content\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.087361 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-catalog-content\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.087438 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-utilities\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.108240 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9psb\" (UniqueName: \"kubernetes.io/projected/b8658e99-6b59-4722-90b2-7a18294722df-kube-api-access-x9psb\") pod \"certified-operators-rkwjn\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.174820 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.400444 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-779vs"] Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.403353 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.405240 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.415568 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-779vs"] Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.490355 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cf50040-7cda-4267-afff-be6e9d288f7a-catalog-content\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.490403 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwmhk\" (UniqueName: \"kubernetes.io/projected/4cf50040-7cda-4267-afff-be6e9d288f7a-kube-api-access-gwmhk\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.490727 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cf50040-7cda-4267-afff-be6e9d288f7a-utilities\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.546150 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkwjn"] Dec 04 13:54:57 crc kubenswrapper[4848]: W1204 13:54:57.547715 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8658e99_6b59_4722_90b2_7a18294722df.slice/crio-067ff03e5f70fefd05a5dfa31456333975ee4fd3401eb425a8271797f5c1625d WatchSource:0}: Error finding container 067ff03e5f70fefd05a5dfa31456333975ee4fd3401eb425a8271797f5c1625d: Status 404 returned error can't find the container with id 067ff03e5f70fefd05a5dfa31456333975ee4fd3401eb425a8271797f5c1625d Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.592036 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cf50040-7cda-4267-afff-be6e9d288f7a-utilities\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.592094 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cf50040-7cda-4267-afff-be6e9d288f7a-catalog-content\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.592118 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwmhk\" (UniqueName: \"kubernetes.io/projected/4cf50040-7cda-4267-afff-be6e9d288f7a-kube-api-access-gwmhk\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.592558 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cf50040-7cda-4267-afff-be6e9d288f7a-utilities\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.592724 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cf50040-7cda-4267-afff-be6e9d288f7a-catalog-content\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.614382 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwmhk\" (UniqueName: \"kubernetes.io/projected/4cf50040-7cda-4267-afff-be6e9d288f7a-kube-api-access-gwmhk\") pod \"redhat-marketplace-779vs\" (UID: \"4cf50040-7cda-4267-afff-be6e9d288f7a\") " pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:57 crc kubenswrapper[4848]: I1204 13:54:57.730737 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:54:58 crc kubenswrapper[4848]: I1204 13:54:58.156034 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-779vs"] Dec 04 13:54:58 crc kubenswrapper[4848]: I1204 13:54:58.289307 4848 generic.go:334] "Generic (PLEG): container finished" podID="b8658e99-6b59-4722-90b2-7a18294722df" containerID="7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557" exitCode=0 Dec 04 13:54:58 crc kubenswrapper[4848]: I1204 13:54:58.289633 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkwjn" event={"ID":"b8658e99-6b59-4722-90b2-7a18294722df","Type":"ContainerDied","Data":"7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557"} Dec 04 13:54:58 crc kubenswrapper[4848]: I1204 13:54:58.289780 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkwjn" event={"ID":"b8658e99-6b59-4722-90b2-7a18294722df","Type":"ContainerStarted","Data":"067ff03e5f70fefd05a5dfa31456333975ee4fd3401eb425a8271797f5c1625d"} Dec 04 13:54:58 crc kubenswrapper[4848]: I1204 13:54:58.300483 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8x6gs" Dec 04 13:54:58 crc kubenswrapper[4848]: I1204 13:54:58.311455 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-779vs" event={"ID":"4cf50040-7cda-4267-afff-be6e9d288f7a","Type":"ContainerStarted","Data":"dbc0ec8ed805319466a1a6bbaa81e838a05753990a490da581f55ab8c24b3253"} Dec 04 13:54:58 crc kubenswrapper[4848]: I1204 13:54:58.364995 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fkdjg"] Dec 04 13:54:58 crc kubenswrapper[4848]: I1204 13:54:58.408495 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74527b90-33e5-4580-bee4-e71afc8a89f7" path="/var/lib/kubelet/pods/74527b90-33e5-4580-bee4-e71afc8a89f7/volumes" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.201478 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q2ntw"] Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.202821 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.204883 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.225527 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q2ntw"] Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.314262 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-utilities\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.314317 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x76x2\" (UniqueName: \"kubernetes.io/projected/455f8f2c-0179-4f4c-b5f7-584f0b091527-kube-api-access-x76x2\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.314442 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-catalog-content\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.317870 4848 generic.go:334] "Generic (PLEG): container finished" podID="4cf50040-7cda-4267-afff-be6e9d288f7a" containerID="a67b3857a9efffb74779871af67ef2a815b1f29b60f8bea7c86230711ea5b768" exitCode=0 Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.317912 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-779vs" event={"ID":"4cf50040-7cda-4267-afff-be6e9d288f7a","Type":"ContainerDied","Data":"a67b3857a9efffb74779871af67ef2a815b1f29b60f8bea7c86230711ea5b768"} Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.319766 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkwjn" event={"ID":"b8658e99-6b59-4722-90b2-7a18294722df","Type":"ContainerStarted","Data":"cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8"} Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.415120 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x76x2\" (UniqueName: \"kubernetes.io/projected/455f8f2c-0179-4f4c-b5f7-584f0b091527-kube-api-access-x76x2\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.415229 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-catalog-content\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.415260 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-utilities\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.415675 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-utilities\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.415871 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-catalog-content\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.433580 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x76x2\" (UniqueName: \"kubernetes.io/projected/455f8f2c-0179-4f4c-b5f7-584f0b091527-kube-api-access-x76x2\") pod \"redhat-operators-q2ntw\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.545118 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.732806 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q2ntw"] Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.808349 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s7plv"] Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.810478 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.814256 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7plv"] Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.816557 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.927471 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkzc9\" (UniqueName: \"kubernetes.io/projected/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-kube-api-access-dkzc9\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.927882 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-utilities\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:54:59 crc kubenswrapper[4848]: I1204 13:54:59.928079 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-catalog-content\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.029845 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-catalog-content\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.029908 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkzc9\" (UniqueName: \"kubernetes.io/projected/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-kube-api-access-dkzc9\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.029944 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-utilities\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.030490 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-catalog-content\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.030524 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-utilities\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.048044 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkzc9\" (UniqueName: \"kubernetes.io/projected/16ba46bb-a73e-41e9-bff8-88b2b43e7a23-kube-api-access-dkzc9\") pod \"community-operators-s7plv\" (UID: \"16ba46bb-a73e-41e9-bff8-88b2b43e7a23\") " pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.125583 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.327838 4848 generic.go:334] "Generic (PLEG): container finished" podID="b8658e99-6b59-4722-90b2-7a18294722df" containerID="cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8" exitCode=0 Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.327999 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkwjn" event={"ID":"b8658e99-6b59-4722-90b2-7a18294722df","Type":"ContainerDied","Data":"cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8"} Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.339031 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2ntw" event={"ID":"455f8f2c-0179-4f4c-b5f7-584f0b091527","Type":"ContainerStarted","Data":"94bcb927ab6f81176eac499f3d8a9af22e9043ff9f7b7f707f418bd1327084d1"} Dec 04 13:55:00 crc kubenswrapper[4848]: I1204 13:55:00.508294 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7plv"] Dec 04 13:55:00 crc kubenswrapper[4848]: W1204 13:55:00.509536 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16ba46bb_a73e_41e9_bff8_88b2b43e7a23.slice/crio-651e59dde23972e0919b6e13ed29a7c0f16d7258e3d6f8ad23a10eed2da998a0 WatchSource:0}: Error finding container 651e59dde23972e0919b6e13ed29a7c0f16d7258e3d6f8ad23a10eed2da998a0: Status 404 returned error can't find the container with id 651e59dde23972e0919b6e13ed29a7c0f16d7258e3d6f8ad23a10eed2da998a0 Dec 04 13:55:01 crc kubenswrapper[4848]: I1204 13:55:01.345159 4848 generic.go:334] "Generic (PLEG): container finished" podID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerID="1a5846ee6fc7c1be56b34bbabec3d727df7b9bee399cebe406712ea65ee7121c" exitCode=0 Dec 04 13:55:01 crc kubenswrapper[4848]: I1204 13:55:01.345206 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2ntw" event={"ID":"455f8f2c-0179-4f4c-b5f7-584f0b091527","Type":"ContainerDied","Data":"1a5846ee6fc7c1be56b34bbabec3d727df7b9bee399cebe406712ea65ee7121c"} Dec 04 13:55:01 crc kubenswrapper[4848]: I1204 13:55:01.347589 4848 generic.go:334] "Generic (PLEG): container finished" podID="16ba46bb-a73e-41e9-bff8-88b2b43e7a23" containerID="a3b3630392fa4ff9127d8c801f0d91bbed1ad0360c7056454bd55e3c8a404e18" exitCode=0 Dec 04 13:55:01 crc kubenswrapper[4848]: I1204 13:55:01.347612 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7plv" event={"ID":"16ba46bb-a73e-41e9-bff8-88b2b43e7a23","Type":"ContainerDied","Data":"a3b3630392fa4ff9127d8c801f0d91bbed1ad0360c7056454bd55e3c8a404e18"} Dec 04 13:55:01 crc kubenswrapper[4848]: I1204 13:55:01.347648 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7plv" event={"ID":"16ba46bb-a73e-41e9-bff8-88b2b43e7a23","Type":"ContainerStarted","Data":"651e59dde23972e0919b6e13ed29a7c0f16d7258e3d6f8ad23a10eed2da998a0"} Dec 04 13:55:05 crc kubenswrapper[4848]: I1204 13:55:05.367973 4848 generic.go:334] "Generic (PLEG): container finished" podID="4cf50040-7cda-4267-afff-be6e9d288f7a" containerID="ecea57f2474a94e86b15cb5c0c8787577a6c62b61fd38a4dd84ae03ceb6688c8" exitCode=0 Dec 04 13:55:05 crc kubenswrapper[4848]: I1204 13:55:05.368066 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-779vs" event={"ID":"4cf50040-7cda-4267-afff-be6e9d288f7a","Type":"ContainerDied","Data":"ecea57f2474a94e86b15cb5c0c8787577a6c62b61fd38a4dd84ae03ceb6688c8"} Dec 04 13:55:05 crc kubenswrapper[4848]: I1204 13:55:05.371247 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkwjn" event={"ID":"b8658e99-6b59-4722-90b2-7a18294722df","Type":"ContainerStarted","Data":"a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29"} Dec 04 13:55:05 crc kubenswrapper[4848]: I1204 13:55:05.373869 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2ntw" event={"ID":"455f8f2c-0179-4f4c-b5f7-584f0b091527","Type":"ContainerStarted","Data":"ed188a535dbf6f2864d97c002ddbb897b41c17a6810bda141862a29724d1936b"} Dec 04 13:55:05 crc kubenswrapper[4848]: I1204 13:55:05.409614 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rkwjn" podStartSLOduration=3.530536939 podStartE2EDuration="9.409594797s" podCreationTimestamp="2025-12-04 13:54:56 +0000 UTC" firstStartedPulling="2025-12-04 13:54:58.3104401 +0000 UTC m=+402.252936628" lastFinishedPulling="2025-12-04 13:55:04.189497948 +0000 UTC m=+408.131994486" observedRunningTime="2025-12-04 13:55:05.407109886 +0000 UTC m=+409.349606414" watchObservedRunningTime="2025-12-04 13:55:05.409594797 +0000 UTC m=+409.352091335" Dec 04 13:55:06 crc kubenswrapper[4848]: I1204 13:55:06.381628 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-779vs" event={"ID":"4cf50040-7cda-4267-afff-be6e9d288f7a","Type":"ContainerStarted","Data":"36bb61faac443a9aec7a816e2b774ed0c685fcff76946d1f39365c15c4a00278"} Dec 04 13:55:06 crc kubenswrapper[4848]: I1204 13:55:06.388010 4848 generic.go:334] "Generic (PLEG): container finished" podID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerID="ed188a535dbf6f2864d97c002ddbb897b41c17a6810bda141862a29724d1936b" exitCode=0 Dec 04 13:55:06 crc kubenswrapper[4848]: I1204 13:55:06.388130 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2ntw" event={"ID":"455f8f2c-0179-4f4c-b5f7-584f0b091527","Type":"ContainerDied","Data":"ed188a535dbf6f2864d97c002ddbb897b41c17a6810bda141862a29724d1936b"} Dec 04 13:55:06 crc kubenswrapper[4848]: I1204 13:55:06.390347 4848 generic.go:334] "Generic (PLEG): container finished" podID="16ba46bb-a73e-41e9-bff8-88b2b43e7a23" containerID="f631c28d1862987de0e368488c16987703cfec829b08a4da2d751acfb3ecb7df" exitCode=0 Dec 04 13:55:06 crc kubenswrapper[4848]: I1204 13:55:06.390643 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7plv" event={"ID":"16ba46bb-a73e-41e9-bff8-88b2b43e7a23","Type":"ContainerDied","Data":"f631c28d1862987de0e368488c16987703cfec829b08a4da2d751acfb3ecb7df"} Dec 04 13:55:06 crc kubenswrapper[4848]: I1204 13:55:06.407455 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-779vs" podStartSLOduration=2.9584940360000003 podStartE2EDuration="9.407436569s" podCreationTimestamp="2025-12-04 13:54:57 +0000 UTC" firstStartedPulling="2025-12-04 13:54:59.319454666 +0000 UTC m=+403.261951194" lastFinishedPulling="2025-12-04 13:55:05.768397199 +0000 UTC m=+409.710893727" observedRunningTime="2025-12-04 13:55:06.404941009 +0000 UTC m=+410.347437537" watchObservedRunningTime="2025-12-04 13:55:06.407436569 +0000 UTC m=+410.349933097" Dec 04 13:55:07 crc kubenswrapper[4848]: I1204 13:55:07.175129 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:55:07 crc kubenswrapper[4848]: I1204 13:55:07.175218 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:55:07 crc kubenswrapper[4848]: I1204 13:55:07.248537 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:55:07 crc kubenswrapper[4848]: I1204 13:55:07.735685 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:55:07 crc kubenswrapper[4848]: I1204 13:55:07.736088 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:55:07 crc kubenswrapper[4848]: I1204 13:55:07.778249 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:55:08 crc kubenswrapper[4848]: I1204 13:55:08.404506 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7plv" event={"ID":"16ba46bb-a73e-41e9-bff8-88b2b43e7a23","Type":"ContainerStarted","Data":"d867788a468226ff14346f91f9979d025b98210a7b2060f253e145dd03c34db0"} Dec 04 13:55:08 crc kubenswrapper[4848]: I1204 13:55:08.406550 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2ntw" event={"ID":"455f8f2c-0179-4f4c-b5f7-584f0b091527","Type":"ContainerStarted","Data":"9c6726243aa9ce482f6bef9410b748d2b9445fc325591a6cf970e2f1dbc83b9b"} Dec 04 13:55:08 crc kubenswrapper[4848]: I1204 13:55:08.426079 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s7plv" podStartSLOduration=4.675873532 podStartE2EDuration="9.426059586s" podCreationTimestamp="2025-12-04 13:54:59 +0000 UTC" firstStartedPulling="2025-12-04 13:55:02.79843581 +0000 UTC m=+406.740932338" lastFinishedPulling="2025-12-04 13:55:07.548621824 +0000 UTC m=+411.491118392" observedRunningTime="2025-12-04 13:55:08.420689914 +0000 UTC m=+412.363186442" watchObservedRunningTime="2025-12-04 13:55:08.426059586 +0000 UTC m=+412.368556114" Dec 04 13:55:08 crc kubenswrapper[4848]: I1204 13:55:08.443978 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q2ntw" podStartSLOduration=3.127584891 podStartE2EDuration="9.443944365s" podCreationTimestamp="2025-12-04 13:54:59 +0000 UTC" firstStartedPulling="2025-12-04 13:55:01.346409658 +0000 UTC m=+405.288906186" lastFinishedPulling="2025-12-04 13:55:07.662769132 +0000 UTC m=+411.605265660" observedRunningTime="2025-12-04 13:55:08.439995488 +0000 UTC m=+412.382492026" watchObservedRunningTime="2025-12-04 13:55:08.443944365 +0000 UTC m=+412.386440893" Dec 04 13:55:09 crc kubenswrapper[4848]: I1204 13:55:09.545989 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:55:09 crc kubenswrapper[4848]: I1204 13:55:09.546323 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:55:10 crc kubenswrapper[4848]: I1204 13:55:10.126259 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:10 crc kubenswrapper[4848]: I1204 13:55:10.126601 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:10 crc kubenswrapper[4848]: I1204 13:55:10.167426 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:10 crc kubenswrapper[4848]: I1204 13:55:10.590970 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q2ntw" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="registry-server" probeResult="failure" output=< Dec 04 13:55:10 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 13:55:10 crc kubenswrapper[4848]: > Dec 04 13:55:14 crc kubenswrapper[4848]: I1204 13:55:14.314640 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:55:14 crc kubenswrapper[4848]: I1204 13:55:14.315142 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:55:14 crc kubenswrapper[4848]: I1204 13:55:14.315183 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:55:14 crc kubenswrapper[4848]: I1204 13:55:14.315583 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"72d765596d3c5a2d39e725591250e92b81042c983d061e6907f508172e90aa2a"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:55:14 crc kubenswrapper[4848]: I1204 13:55:14.315635 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://72d765596d3c5a2d39e725591250e92b81042c983d061e6907f508172e90aa2a" gracePeriod=600 Dec 04 13:55:15 crc kubenswrapper[4848]: I1204 13:55:15.444084 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="72d765596d3c5a2d39e725591250e92b81042c983d061e6907f508172e90aa2a" exitCode=0 Dec 04 13:55:15 crc kubenswrapper[4848]: I1204 13:55:15.444170 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"72d765596d3c5a2d39e725591250e92b81042c983d061e6907f508172e90aa2a"} Dec 04 13:55:15 crc kubenswrapper[4848]: I1204 13:55:15.444410 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"c3cebfe15e9ee005602e42c9fe3958ff03b1e9bc1b525474da30cb9fd280b5af"} Dec 04 13:55:15 crc kubenswrapper[4848]: I1204 13:55:15.444448 4848 scope.go:117] "RemoveContainer" containerID="305492fc53854c76e85a3f806ef231d4c6a16619ab798d0854c5a6da090106b1" Dec 04 13:55:17 crc kubenswrapper[4848]: I1204 13:55:17.213701 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 13:55:17 crc kubenswrapper[4848]: I1204 13:55:17.793054 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-779vs" Dec 04 13:55:19 crc kubenswrapper[4848]: I1204 13:55:19.605990 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:55:19 crc kubenswrapper[4848]: I1204 13:55:19.651785 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 13:55:20 crc kubenswrapper[4848]: I1204 13:55:20.196917 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s7plv" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.410475 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" podUID="9b739fda-b42c-49be-99dc-4ff123bb8cb7" containerName="registry" containerID="cri-o://605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594" gracePeriod=30 Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.763873 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.846669 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzq79\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-kube-api-access-kzq79\") pod \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.846715 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b739fda-b42c-49be-99dc-4ff123bb8cb7-ca-trust-extracted\") pod \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.846732 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-certificates\") pod \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.847780 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9b739fda-b42c-49be-99dc-4ff123bb8cb7" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.847880 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.847975 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-bound-sa-token\") pod \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.848033 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-tls\") pod \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.848073 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b739fda-b42c-49be-99dc-4ff123bb8cb7-installation-pull-secrets\") pod \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.848105 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-trusted-ca\") pod \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\" (UID: \"9b739fda-b42c-49be-99dc-4ff123bb8cb7\") " Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.848302 4848 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.848625 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9b739fda-b42c-49be-99dc-4ff123bb8cb7" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.854153 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9b739fda-b42c-49be-99dc-4ff123bb8cb7" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.854489 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b739fda-b42c-49be-99dc-4ff123bb8cb7-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9b739fda-b42c-49be-99dc-4ff123bb8cb7" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.854804 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-kube-api-access-kzq79" (OuterVolumeSpecName: "kube-api-access-kzq79") pod "9b739fda-b42c-49be-99dc-4ff123bb8cb7" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7"). InnerVolumeSpecName "kube-api-access-kzq79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.856403 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9b739fda-b42c-49be-99dc-4ff123bb8cb7" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.858265 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9b739fda-b42c-49be-99dc-4ff123bb8cb7" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.884525 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b739fda-b42c-49be-99dc-4ff123bb8cb7-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9b739fda-b42c-49be-99dc-4ff123bb8cb7" (UID: "9b739fda-b42c-49be-99dc-4ff123bb8cb7"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.949642 4848 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.949671 4848 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.949680 4848 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b739fda-b42c-49be-99dc-4ff123bb8cb7-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.949690 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b739fda-b42c-49be-99dc-4ff123bb8cb7-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.949701 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzq79\" (UniqueName: \"kubernetes.io/projected/9b739fda-b42c-49be-99dc-4ff123bb8cb7-kube-api-access-kzq79\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:23 crc kubenswrapper[4848]: I1204 13:55:23.949709 4848 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b739fda-b42c-49be-99dc-4ff123bb8cb7-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.491236 4848 generic.go:334] "Generic (PLEG): container finished" podID="9b739fda-b42c-49be-99dc-4ff123bb8cb7" containerID="605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594" exitCode=0 Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.491285 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" event={"ID":"9b739fda-b42c-49be-99dc-4ff123bb8cb7","Type":"ContainerDied","Data":"605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594"} Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.491301 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.491330 4848 scope.go:117] "RemoveContainer" containerID="605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594" Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.491319 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fkdjg" event={"ID":"9b739fda-b42c-49be-99dc-4ff123bb8cb7","Type":"ContainerDied","Data":"8fabbc7a3fdb48c79c566be5061ab4219e8fc8fffa3dee3dd2a337aaeafa431b"} Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.508824 4848 scope.go:117] "RemoveContainer" containerID="605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594" Dec 04 13:55:24 crc kubenswrapper[4848]: E1204 13:55:24.509353 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594\": container with ID starting with 605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594 not found: ID does not exist" containerID="605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594" Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.509383 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594"} err="failed to get container status \"605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594\": rpc error: code = NotFound desc = could not find container \"605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594\": container with ID starting with 605ca3028590a8a3008b976607f9d8cbc472dc03402e1d34e92b4b92fdc11594 not found: ID does not exist" Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.513239 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fkdjg"] Dec 04 13:55:24 crc kubenswrapper[4848]: I1204 13:55:24.518141 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fkdjg"] Dec 04 13:55:26 crc kubenswrapper[4848]: I1204 13:55:26.408228 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b739fda-b42c-49be-99dc-4ff123bb8cb7" path="/var/lib/kubelet/pods/9b739fda-b42c-49be-99dc-4ff123bb8cb7/volumes" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.064894 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg"] Dec 04 13:55:27 crc kubenswrapper[4848]: E1204 13:55:27.065122 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b739fda-b42c-49be-99dc-4ff123bb8cb7" containerName="registry" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.065134 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b739fda-b42c-49be-99dc-4ff123bb8cb7" containerName="registry" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.065253 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b739fda-b42c-49be-99dc-4ff123bb8cb7" containerName="registry" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.065666 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.069054 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.069115 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.069054 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.069586 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.072922 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.083923 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg"] Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.186319 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkzw9\" (UniqueName: \"kubernetes.io/projected/356e5ce4-4154-4e6f-b8cf-08c0244199e7-kube-api-access-kkzw9\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.186419 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/356e5ce4-4154-4e6f-b8cf-08c0244199e7-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.186462 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/356e5ce4-4154-4e6f-b8cf-08c0244199e7-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.288157 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkzw9\" (UniqueName: \"kubernetes.io/projected/356e5ce4-4154-4e6f-b8cf-08c0244199e7-kube-api-access-kkzw9\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.288272 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/356e5ce4-4154-4e6f-b8cf-08c0244199e7-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.288313 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/356e5ce4-4154-4e6f-b8cf-08c0244199e7-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.289710 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/356e5ce4-4154-4e6f-b8cf-08c0244199e7-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.296040 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/356e5ce4-4154-4e6f-b8cf-08c0244199e7-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.305106 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkzw9\" (UniqueName: \"kubernetes.io/projected/356e5ce4-4154-4e6f-b8cf-08c0244199e7-kube-api-access-kkzw9\") pod \"cluster-monitoring-operator-6d5b84845-v7smg\" (UID: \"356e5ce4-4154-4e6f-b8cf-08c0244199e7\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.385066 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" Dec 04 13:55:27 crc kubenswrapper[4848]: I1204 13:55:27.594806 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg"] Dec 04 13:55:27 crc kubenswrapper[4848]: W1204 13:55:27.597821 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356e5ce4_4154_4e6f_b8cf_08c0244199e7.slice/crio-8666ed46bc6e7a63fa43984252fc4cfb6bf4b842079f921ee72d723ac40cdcff WatchSource:0}: Error finding container 8666ed46bc6e7a63fa43984252fc4cfb6bf4b842079f921ee72d723ac40cdcff: Status 404 returned error can't find the container with id 8666ed46bc6e7a63fa43984252fc4cfb6bf4b842079f921ee72d723ac40cdcff Dec 04 13:55:28 crc kubenswrapper[4848]: I1204 13:55:28.513023 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" event={"ID":"356e5ce4-4154-4e6f-b8cf-08c0244199e7","Type":"ContainerStarted","Data":"8666ed46bc6e7a63fa43984252fc4cfb6bf4b842079f921ee72d723ac40cdcff"} Dec 04 13:55:29 crc kubenswrapper[4848]: I1204 13:55:29.520495 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" event={"ID":"356e5ce4-4154-4e6f-b8cf-08c0244199e7","Type":"ContainerStarted","Data":"8ab564565931fc070a84f961b37326f09222b7a2b3fb0d4ea5437772c98d45ee"} Dec 04 13:55:29 crc kubenswrapper[4848]: I1204 13:55:29.540918 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-v7smg" podStartSLOduration=0.844827939 podStartE2EDuration="2.540895562s" podCreationTimestamp="2025-12-04 13:55:27 +0000 UTC" firstStartedPulling="2025-12-04 13:55:27.599918527 +0000 UTC m=+431.542415065" lastFinishedPulling="2025-12-04 13:55:29.29598616 +0000 UTC m=+433.238482688" observedRunningTime="2025-12-04 13:55:29.539289012 +0000 UTC m=+433.481785540" watchObservedRunningTime="2025-12-04 13:55:29.540895562 +0000 UTC m=+433.483392090" Dec 04 13:55:29 crc kubenswrapper[4848]: I1204 13:55:29.860682 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d"] Dec 04 13:55:29 crc kubenswrapper[4848]: I1204 13:55:29.861374 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" Dec 04 13:55:29 crc kubenswrapper[4848]: I1204 13:55:29.863943 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Dec 04 13:55:29 crc kubenswrapper[4848]: I1204 13:55:29.866696 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-6gbpm" Dec 04 13:55:29 crc kubenswrapper[4848]: I1204 13:55:29.870627 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d"] Dec 04 13:55:29 crc kubenswrapper[4848]: I1204 13:55:29.916404 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/26714bdc-965a-4b7b-b4b4-53d1555f7bd4-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-6985d\" (UID: \"26714bdc-965a-4b7b-b4b4-53d1555f7bd4\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" Dec 04 13:55:30 crc kubenswrapper[4848]: I1204 13:55:30.018106 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/26714bdc-965a-4b7b-b4b4-53d1555f7bd4-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-6985d\" (UID: \"26714bdc-965a-4b7b-b4b4-53d1555f7bd4\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" Dec 04 13:55:30 crc kubenswrapper[4848]: E1204 13:55:30.018322 4848 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Dec 04 13:55:30 crc kubenswrapper[4848]: E1204 13:55:30.018406 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26714bdc-965a-4b7b-b4b4-53d1555f7bd4-tls-certificates podName:26714bdc-965a-4b7b-b4b4-53d1555f7bd4 nodeName:}" failed. No retries permitted until 2025-12-04 13:55:30.518383853 +0000 UTC m=+434.460880401 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/26714bdc-965a-4b7b-b4b4-53d1555f7bd4-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-6985d" (UID: "26714bdc-965a-4b7b-b4b4-53d1555f7bd4") : secret "prometheus-operator-admission-webhook-tls" not found Dec 04 13:55:30 crc kubenswrapper[4848]: I1204 13:55:30.524510 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/26714bdc-965a-4b7b-b4b4-53d1555f7bd4-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-6985d\" (UID: \"26714bdc-965a-4b7b-b4b4-53d1555f7bd4\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" Dec 04 13:55:30 crc kubenswrapper[4848]: I1204 13:55:30.539216 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/26714bdc-965a-4b7b-b4b4-53d1555f7bd4-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-6985d\" (UID: \"26714bdc-965a-4b7b-b4b4-53d1555f7bd4\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" Dec 04 13:55:30 crc kubenswrapper[4848]: I1204 13:55:30.776866 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" Dec 04 13:55:31 crc kubenswrapper[4848]: I1204 13:55:31.260131 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d"] Dec 04 13:55:31 crc kubenswrapper[4848]: W1204 13:55:31.274340 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26714bdc_965a_4b7b_b4b4_53d1555f7bd4.slice/crio-e1dcd3a5192b86b4d8b481984d1eab6760f78986bd7566c31fb6906143901e2c WatchSource:0}: Error finding container e1dcd3a5192b86b4d8b481984d1eab6760f78986bd7566c31fb6906143901e2c: Status 404 returned error can't find the container with id e1dcd3a5192b86b4d8b481984d1eab6760f78986bd7566c31fb6906143901e2c Dec 04 13:55:31 crc kubenswrapper[4848]: I1204 13:55:31.533349 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" event={"ID":"26714bdc-965a-4b7b-b4b4-53d1555f7bd4","Type":"ContainerStarted","Data":"e1dcd3a5192b86b4d8b481984d1eab6760f78986bd7566c31fb6906143901e2c"} Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.546050 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" event={"ID":"26714bdc-965a-4b7b-b4b4-53d1555f7bd4","Type":"ContainerStarted","Data":"dc0f93f8ed34fddb611c16b3c3ca51d01a9890e9ad4d8840ad1b5f881dba6200"} Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.546584 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.555102 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.563669 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-6985d" podStartSLOduration=3.358919998 podStartE2EDuration="4.563650321s" podCreationTimestamp="2025-12-04 13:55:29 +0000 UTC" firstStartedPulling="2025-12-04 13:55:31.275752804 +0000 UTC m=+435.218249332" lastFinishedPulling="2025-12-04 13:55:32.480483117 +0000 UTC m=+436.422979655" observedRunningTime="2025-12-04 13:55:33.56074797 +0000 UTC m=+437.503244508" watchObservedRunningTime="2025-12-04 13:55:33.563650321 +0000 UTC m=+437.506146859" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.917647 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-gzbq4"] Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.918432 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.920352 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.920445 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.920520 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.921192 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-hbh5s" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.928358 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-gzbq4"] Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.970641 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.970748 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.970790 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2e0da298-378f-4a43-8376-5d82ee2ed0e4-metrics-client-ca\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:33 crc kubenswrapper[4848]: I1204 13:55:33.970827 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdbzw\" (UniqueName: \"kubernetes.io/projected/2e0da298-378f-4a43-8376-5d82ee2ed0e4-kube-api-access-xdbzw\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.072674 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.072771 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2e0da298-378f-4a43-8376-5d82ee2ed0e4-metrics-client-ca\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: E1204 13:55:34.072809 4848 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-tls: secret "prometheus-operator-tls" not found Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.072824 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdbzw\" (UniqueName: \"kubernetes.io/projected/2e0da298-378f-4a43-8376-5d82ee2ed0e4-kube-api-access-xdbzw\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: E1204 13:55:34.072877 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-tls podName:2e0da298-378f-4a43-8376-5d82ee2ed0e4 nodeName:}" failed. No retries permitted until 2025-12-04 13:55:34.572857029 +0000 UTC m=+438.515353567 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-operator-tls" (UniqueName: "kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-tls") pod "prometheus-operator-db54df47d-gzbq4" (UID: "2e0da298-378f-4a43-8376-5d82ee2ed0e4") : secret "prometheus-operator-tls" not found Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.072903 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.074546 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2e0da298-378f-4a43-8376-5d82ee2ed0e4-metrics-client-ca\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.080595 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.101633 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdbzw\" (UniqueName: \"kubernetes.io/projected/2e0da298-378f-4a43-8376-5d82ee2ed0e4-kube-api-access-xdbzw\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.581594 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.590374 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/2e0da298-378f-4a43-8376-5d82ee2ed0e4-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-gzbq4\" (UID: \"2e0da298-378f-4a43-8376-5d82ee2ed0e4\") " pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:34 crc kubenswrapper[4848]: I1204 13:55:34.839375 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" Dec 04 13:55:35 crc kubenswrapper[4848]: I1204 13:55:35.269926 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-gzbq4"] Dec 04 13:55:35 crc kubenswrapper[4848]: I1204 13:55:35.561758 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" event={"ID":"2e0da298-378f-4a43-8376-5d82ee2ed0e4","Type":"ContainerStarted","Data":"550e1eee1cc49542fa07d47c6b68c96e49ca6545f380f22da95e14024ef29a68"} Dec 04 13:55:37 crc kubenswrapper[4848]: I1204 13:55:37.574139 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" event={"ID":"2e0da298-378f-4a43-8376-5d82ee2ed0e4","Type":"ContainerStarted","Data":"b17f1363b5b6d7ab05001f3eae0ba223ffc534f9d39c63761eb720be16d4ba35"} Dec 04 13:55:37 crc kubenswrapper[4848]: I1204 13:55:37.574547 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" event={"ID":"2e0da298-378f-4a43-8376-5d82ee2ed0e4","Type":"ContainerStarted","Data":"05cdf4ad1b04f5315ac3dd9701987e1e90f1a1a5ae760df00b524f284809d20a"} Dec 04 13:55:37 crc kubenswrapper[4848]: I1204 13:55:37.597655 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-gzbq4" podStartSLOduration=3.269682903 podStartE2EDuration="4.597621415s" podCreationTimestamp="2025-12-04 13:55:33 +0000 UTC" firstStartedPulling="2025-12-04 13:55:35.277052538 +0000 UTC m=+439.219549066" lastFinishedPulling="2025-12-04 13:55:36.60499105 +0000 UTC m=+440.547487578" observedRunningTime="2025-12-04 13:55:37.59212695 +0000 UTC m=+441.534623518" watchObservedRunningTime="2025-12-04 13:55:37.597621415 +0000 UTC m=+441.540117983" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.261271 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq"] Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.262288 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.265119 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.265324 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-jvq25" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.265898 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.278621 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq"] Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.287015 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s"] Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.288239 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.294371 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.294563 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-2m674" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.294869 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.296268 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.300338 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s"] Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.331537 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-h2v59"] Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.332760 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.336323 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.336450 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-9gwnm" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.336525 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364095 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364139 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bbzf\" (UniqueName: \"kubernetes.io/projected/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-api-access-2bbzf\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364168 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmfbd\" (UniqueName: \"kubernetes.io/projected/008006cc-05ca-43ca-a20c-84392033b712-kube-api-access-mmfbd\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364189 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-root\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364212 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-textfile\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364268 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364372 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/008006cc-05ca-43ca-a20c-84392033b712-metrics-client-ca\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364406 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-wtmp\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364464 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f91fa164-b3ba-437c-b2cb-4e40a4857371-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364523 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364551 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364608 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-tls\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364636 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/f91fa164-b3ba-437c-b2cb-4e40a4857371-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364699 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-sys\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364729 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364754 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9q9b\" (UniqueName: \"kubernetes.io/projected/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-kube-api-access-n9q9b\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364777 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.364797 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466480 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9q9b\" (UniqueName: \"kubernetes.io/projected/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-kube-api-access-n9q9b\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466525 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466542 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466563 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466581 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bbzf\" (UniqueName: \"kubernetes.io/projected/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-api-access-2bbzf\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466611 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmfbd\" (UniqueName: \"kubernetes.io/projected/008006cc-05ca-43ca-a20c-84392033b712-kube-api-access-mmfbd\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466630 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-root\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466649 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-textfile\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466667 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466697 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/008006cc-05ca-43ca-a20c-84392033b712-metrics-client-ca\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466721 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-wtmp\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466740 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f91fa164-b3ba-437c-b2cb-4e40a4857371-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466766 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466801 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466819 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-tls\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466837 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/f91fa164-b3ba-437c-b2cb-4e40a4857371-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466862 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-sys\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.466879 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.467923 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.468182 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-root\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.468507 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-textfile\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.469257 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.470151 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/f91fa164-b3ba-437c-b2cb-4e40a4857371-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.470220 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-sys\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.470629 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/008006cc-05ca-43ca-a20c-84392033b712-metrics-client-ca\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.470780 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f91fa164-b3ba-437c-b2cb-4e40a4857371-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.470801 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-wtmp\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.472527 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.472770 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.473182 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/008006cc-05ca-43ca-a20c-84392033b712-node-exporter-tls\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.473269 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.473571 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.474060 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.522292 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bbzf\" (UniqueName: \"kubernetes.io/projected/f91fa164-b3ba-437c-b2cb-4e40a4857371-kube-api-access-2bbzf\") pod \"kube-state-metrics-777cb5bd5d-lr42s\" (UID: \"f91fa164-b3ba-437c-b2cb-4e40a4857371\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.524746 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9q9b\" (UniqueName: \"kubernetes.io/projected/07f4507c-7d5f-4ab9-ab02-b47c5ad64943-kube-api-access-n9q9b\") pod \"openshift-state-metrics-566fddb674-vqzfq\" (UID: \"07f4507c-7d5f-4ab9-ab02-b47c5ad64943\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.524825 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmfbd\" (UniqueName: \"kubernetes.io/projected/008006cc-05ca-43ca-a20c-84392033b712-kube-api-access-mmfbd\") pod \"node-exporter-h2v59\" (UID: \"008006cc-05ca-43ca-a20c-84392033b712\") " pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.576533 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.607994 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" Dec 04 13:55:39 crc kubenswrapper[4848]: I1204 13:55:39.648809 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-h2v59" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.053847 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq"] Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.141874 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s"] Dec 04 13:55:40 crc kubenswrapper[4848]: W1204 13:55:40.149637 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf91fa164_b3ba_437c_b2cb_4e40a4857371.slice/crio-385f1ef681a981828857b0ebd382f1436ea0e15ad63f7412c94a31a1158fea94 WatchSource:0}: Error finding container 385f1ef681a981828857b0ebd382f1436ea0e15ad63f7412c94a31a1158fea94: Status 404 returned error can't find the container with id 385f1ef681a981828857b0ebd382f1436ea0e15ad63f7412c94a31a1158fea94 Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.319261 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.321344 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.324168 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.324215 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.324736 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.324999 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.325054 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.325861 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-r58wn" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.325906 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.329194 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.330665 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.353819 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.377981 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378030 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-config-volume\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378058 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378217 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-config-out\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378294 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnjj4\" (UniqueName: \"kubernetes.io/projected/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-kube-api-access-fnjj4\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378343 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-web-config\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378365 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-tls-assets\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378418 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378446 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378542 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378573 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.378596 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480085 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnjj4\" (UniqueName: \"kubernetes.io/projected/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-kube-api-access-fnjj4\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480164 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-web-config\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480192 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-tls-assets\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480254 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480303 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480383 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480411 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480460 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480495 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480546 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-config-volume\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480574 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.480635 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-config-out\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.481536 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.481607 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.482460 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.488543 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-web-config\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.488571 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-config-volume\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.488652 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.488941 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-tls-assets\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.488985 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.489361 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.492334 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-config-out\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.492507 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.498756 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnjj4\" (UniqueName: \"kubernetes.io/projected/f8fc10b8-4f17-4ca9-affc-4578ff6fac06-kube-api-access-fnjj4\") pod \"alertmanager-main-0\" (UID: \"f8fc10b8-4f17-4ca9-affc-4578ff6fac06\") " pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.590140 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-h2v59" event={"ID":"008006cc-05ca-43ca-a20c-84392033b712","Type":"ContainerStarted","Data":"66651a74f2345d39b29b942b0ef4c0126063ee10824187fc08cd7e9ced92d127"} Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.591140 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" event={"ID":"07f4507c-7d5f-4ab9-ab02-b47c5ad64943","Type":"ContainerStarted","Data":"08dd858719a3782e9e84025d73046ee6d641a5e9f205ca339fbb15824c6649cc"} Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.591831 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" event={"ID":"f91fa164-b3ba-437c-b2cb-4e40a4857371","Type":"ContainerStarted","Data":"385f1ef681a981828857b0ebd382f1436ea0e15ad63f7412c94a31a1158fea94"} Dec 04 13:55:40 crc kubenswrapper[4848]: I1204 13:55:40.640408 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.118135 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 04 13:55:41 crc kubenswrapper[4848]: W1204 13:55:41.124488 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8fc10b8_4f17_4ca9_affc_4578ff6fac06.slice/crio-fb3ba4f723ba15d9eccac109020f0f51001f3e4e4a3dff7c16e620b1dbf8bb25 WatchSource:0}: Error finding container fb3ba4f723ba15d9eccac109020f0f51001f3e4e4a3dff7c16e620b1dbf8bb25: Status 404 returned error can't find the container with id fb3ba4f723ba15d9eccac109020f0f51001f3e4e4a3dff7c16e620b1dbf8bb25 Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.352756 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-779569c9f6-hdxhl"] Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.354380 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.357176 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-f7mdb" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.357230 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.357335 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.357503 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.357585 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-4l54cao0k1r5" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.358073 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.361717 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.376341 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-779569c9f6-hdxhl"] Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.401799 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-tls\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.401850 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbbnd\" (UniqueName: \"kubernetes.io/projected/568712c9-5113-4808-a56f-f58acf42eb47-kube-api-access-bbbnd\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.401874 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.401914 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-grpc-tls\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.401932 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.401971 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.401990 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/568712c9-5113-4808-a56f-f58acf42eb47-metrics-client-ca\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.402017 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.503021 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.503083 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/568712c9-5113-4808-a56f-f58acf42eb47-metrics-client-ca\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.503115 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.503253 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-tls\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.503292 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbbnd\" (UniqueName: \"kubernetes.io/projected/568712c9-5113-4808-a56f-f58acf42eb47-kube-api-access-bbbnd\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.503331 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.503382 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-grpc-tls\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.503407 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.504927 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/568712c9-5113-4808-a56f-f58acf42eb47-metrics-client-ca\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.507805 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.511309 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.511447 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.511720 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-grpc-tls\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.511765 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.517492 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/568712c9-5113-4808-a56f-f58acf42eb47-secret-thanos-querier-tls\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.523003 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbbnd\" (UniqueName: \"kubernetes.io/projected/568712c9-5113-4808-a56f-f58acf42eb47-kube-api-access-bbbnd\") pod \"thanos-querier-779569c9f6-hdxhl\" (UID: \"568712c9-5113-4808-a56f-f58acf42eb47\") " pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.601072 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8fc10b8-4f17-4ca9-affc-4578ff6fac06","Type":"ContainerStarted","Data":"fb3ba4f723ba15d9eccac109020f0f51001f3e4e4a3dff7c16e620b1dbf8bb25"} Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.603672 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" event={"ID":"07f4507c-7d5f-4ab9-ab02-b47c5ad64943","Type":"ContainerStarted","Data":"05cdb64197ea8c6e7c40d112ecdd8caecc9f90571f9a7fb58cb3f75cbaec6bd4"} Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.603716 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" event={"ID":"07f4507c-7d5f-4ab9-ab02-b47c5ad64943","Type":"ContainerStarted","Data":"0ad270f3a273a596ddc670891d8b7f341ccae7e0ec11fc35717e6554d629f959"} Dec 04 13:55:41 crc kubenswrapper[4848]: I1204 13:55:41.675553 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:42 crc kubenswrapper[4848]: W1204 13:55:42.399559 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod568712c9_5113_4808_a56f_f58acf42eb47.slice/crio-2761c44e60d75961b8d757a0b3c66c895b9c968a812ea3ed7bd16df5220f4604 WatchSource:0}: Error finding container 2761c44e60d75961b8d757a0b3c66c895b9c968a812ea3ed7bd16df5220f4604: Status 404 returned error can't find the container with id 2761c44e60d75961b8d757a0b3c66c895b9c968a812ea3ed7bd16df5220f4604 Dec 04 13:55:42 crc kubenswrapper[4848]: I1204 13:55:42.402268 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-779569c9f6-hdxhl"] Dec 04 13:55:42 crc kubenswrapper[4848]: I1204 13:55:42.613698 4848 generic.go:334] "Generic (PLEG): container finished" podID="008006cc-05ca-43ca-a20c-84392033b712" containerID="c14d7a592ec36371d28ca2d3fb8ce43bb3c9371eb67f777dbb86f5590b6288f5" exitCode=0 Dec 04 13:55:42 crc kubenswrapper[4848]: I1204 13:55:42.613788 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-h2v59" event={"ID":"008006cc-05ca-43ca-a20c-84392033b712","Type":"ContainerDied","Data":"c14d7a592ec36371d28ca2d3fb8ce43bb3c9371eb67f777dbb86f5590b6288f5"} Dec 04 13:55:42 crc kubenswrapper[4848]: I1204 13:55:42.617564 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" event={"ID":"568712c9-5113-4808-a56f-f58acf42eb47","Type":"ContainerStarted","Data":"2761c44e60d75961b8d757a0b3c66c895b9c968a812ea3ed7bd16df5220f4604"} Dec 04 13:55:42 crc kubenswrapper[4848]: I1204 13:55:42.632551 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" event={"ID":"f91fa164-b3ba-437c-b2cb-4e40a4857371","Type":"ContainerStarted","Data":"be2d0878259bfa5c5c66547dc75a6007b130cb278fced02028567e5c3159521e"} Dec 04 13:55:42 crc kubenswrapper[4848]: I1204 13:55:42.632612 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" event={"ID":"f91fa164-b3ba-437c-b2cb-4e40a4857371","Type":"ContainerStarted","Data":"e20c9be6d28c0fe0bdce48d1a9211d74109f87fc61da4d16cef30452e4367bc5"} Dec 04 13:55:42 crc kubenswrapper[4848]: I1204 13:55:42.632626 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" event={"ID":"f91fa164-b3ba-437c-b2cb-4e40a4857371","Type":"ContainerStarted","Data":"1238bbe563d752783cf91dc88b52e9a11650c45292ff416fb341a84e6c1ebdae"} Dec 04 13:55:42 crc kubenswrapper[4848]: I1204 13:55:42.657269 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-lr42s" podStartSLOduration=1.796114355 podStartE2EDuration="3.657241563s" podCreationTimestamp="2025-12-04 13:55:39 +0000 UTC" firstStartedPulling="2025-12-04 13:55:40.151770775 +0000 UTC m=+444.094267293" lastFinishedPulling="2025-12-04 13:55:42.012897973 +0000 UTC m=+445.955394501" observedRunningTime="2025-12-04 13:55:42.652203179 +0000 UTC m=+446.594699707" watchObservedRunningTime="2025-12-04 13:55:42.657241563 +0000 UTC m=+446.599738131" Dec 04 13:55:43 crc kubenswrapper[4848]: I1204 13:55:43.639515 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-h2v59" event={"ID":"008006cc-05ca-43ca-a20c-84392033b712","Type":"ContainerStarted","Data":"7631cdbc322f6069938280ed58081e6a3fec63a482264360f01f2cdcf23e2f71"} Dec 04 13:55:43 crc kubenswrapper[4848]: I1204 13:55:43.640181 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-h2v59" event={"ID":"008006cc-05ca-43ca-a20c-84392033b712","Type":"ContainerStarted","Data":"1051ace35a72eb15a060cb5aa22eaaadc37a50af139c554736f6f1c0d566f1d6"} Dec 04 13:55:43 crc kubenswrapper[4848]: I1204 13:55:43.643094 4848 generic.go:334] "Generic (PLEG): container finished" podID="f8fc10b8-4f17-4ca9-affc-4578ff6fac06" containerID="b1abc1ac2bac6d001af1400faffdccb7c9c3fad4a9824c03a041e076134429cd" exitCode=0 Dec 04 13:55:43 crc kubenswrapper[4848]: I1204 13:55:43.643937 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8fc10b8-4f17-4ca9-affc-4578ff6fac06","Type":"ContainerDied","Data":"b1abc1ac2bac6d001af1400faffdccb7c9c3fad4a9824c03a041e076134429cd"} Dec 04 13:55:43 crc kubenswrapper[4848]: I1204 13:55:43.648822 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" event={"ID":"07f4507c-7d5f-4ab9-ab02-b47c5ad64943","Type":"ContainerStarted","Data":"f05a5db576cbc42f3bc9b6c6a182a2d4394bd181b4ff614ee9f27405bdbc73d4"} Dec 04 13:55:43 crc kubenswrapper[4848]: I1204 13:55:43.666798 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-h2v59" podStartSLOduration=2.381101681 podStartE2EDuration="4.666779802s" podCreationTimestamp="2025-12-04 13:55:39 +0000 UTC" firstStartedPulling="2025-12-04 13:55:39.671974188 +0000 UTC m=+443.614470716" lastFinishedPulling="2025-12-04 13:55:41.957652309 +0000 UTC m=+445.900148837" observedRunningTime="2025-12-04 13:55:43.656908539 +0000 UTC m=+447.599405067" watchObservedRunningTime="2025-12-04 13:55:43.666779802 +0000 UTC m=+447.609276330" Dec 04 13:55:43 crc kubenswrapper[4848]: I1204 13:55:43.686856 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-vqzfq" podStartSLOduration=2.699733988 podStartE2EDuration="4.686837023s" podCreationTimestamp="2025-12-04 13:55:39 +0000 UTC" firstStartedPulling="2025-12-04 13:55:41.214645891 +0000 UTC m=+445.157142419" lastFinishedPulling="2025-12-04 13:55:43.201748926 +0000 UTC m=+447.144245454" observedRunningTime="2025-12-04 13:55:43.679844432 +0000 UTC m=+447.622341000" watchObservedRunningTime="2025-12-04 13:55:43.686837023 +0000 UTC m=+447.629333551" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.093590 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-69f4c7668-cqf5s"] Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.094360 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.114080 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-69f4c7668-cqf5s"] Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.142319 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnqwc\" (UniqueName: \"kubernetes.io/projected/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-kube-api-access-dnqwc\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.142385 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-trusted-ca-bundle\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.142426 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-config\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.142456 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-oauth-serving-cert\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.142486 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-serving-cert\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.142570 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-oauth-config\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.142608 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-service-ca\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.243434 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-oauth-config\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.243489 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-service-ca\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.243527 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnqwc\" (UniqueName: \"kubernetes.io/projected/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-kube-api-access-dnqwc\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.243555 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-trusted-ca-bundle\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.243580 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-config\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.243599 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-oauth-serving-cert\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.243620 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-serving-cert\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.244801 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-oauth-serving-cert\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.244932 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-config\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.245202 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-trusted-ca-bundle\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.246000 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-service-ca\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.248065 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-serving-cert\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.248228 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-oauth-config\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.261094 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnqwc\" (UniqueName: \"kubernetes.io/projected/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-kube-api-access-dnqwc\") pod \"console-69f4c7668-cqf5s\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.409919 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.585321 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-848dcdfb99-f4gmr"] Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.586306 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.589692 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.589871 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.590008 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dakhfon14eto8" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.591298 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.591427 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-nmrzl" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.591480 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.593677 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-848dcdfb99-f4gmr"] Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.648877 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/120107aa-e225-488c-a258-d4d19ccc7ce7-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.648929 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/120107aa-e225-488c-a258-d4d19ccc7ce7-metrics-server-audit-profiles\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.648971 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-client-ca-bundle\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.649001 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-secret-metrics-server-tls\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.649026 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67pfv\" (UniqueName: \"kubernetes.io/projected/120107aa-e225-488c-a258-d4d19ccc7ce7-kube-api-access-67pfv\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.649049 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/120107aa-e225-488c-a258-d4d19ccc7ce7-audit-log\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.649096 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-secret-metrics-client-certs\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.675454 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-69f4c7668-cqf5s"] Dec 04 13:55:44 crc kubenswrapper[4848]: W1204 13:55:44.679157 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02eac971_3e5f_4d8b_baab_5a1e68f0ac7b.slice/crio-e41edb87d9910c48ad45d7f3072a439b85da7c01ccd368e4da0bb6c30dcaec62 WatchSource:0}: Error finding container e41edb87d9910c48ad45d7f3072a439b85da7c01ccd368e4da0bb6c30dcaec62: Status 404 returned error can't find the container with id e41edb87d9910c48ad45d7f3072a439b85da7c01ccd368e4da0bb6c30dcaec62 Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.749897 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-secret-metrics-client-certs\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.749993 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/120107aa-e225-488c-a258-d4d19ccc7ce7-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.750059 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/120107aa-e225-488c-a258-d4d19ccc7ce7-metrics-server-audit-profiles\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.750089 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-client-ca-bundle\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.750175 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-secret-metrics-server-tls\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.750219 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67pfv\" (UniqueName: \"kubernetes.io/projected/120107aa-e225-488c-a258-d4d19ccc7ce7-kube-api-access-67pfv\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.750257 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/120107aa-e225-488c-a258-d4d19ccc7ce7-audit-log\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.751547 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/120107aa-e225-488c-a258-d4d19ccc7ce7-audit-log\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.752144 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/120107aa-e225-488c-a258-d4d19ccc7ce7-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.752845 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/120107aa-e225-488c-a258-d4d19ccc7ce7-metrics-server-audit-profiles\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.757154 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-client-ca-bundle\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.757776 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-secret-metrics-server-tls\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.761970 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/120107aa-e225-488c-a258-d4d19ccc7ce7-secret-metrics-client-certs\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.770877 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67pfv\" (UniqueName: \"kubernetes.io/projected/120107aa-e225-488c-a258-d4d19ccc7ce7-kube-api-access-67pfv\") pod \"metrics-server-848dcdfb99-f4gmr\" (UID: \"120107aa-e225-488c-a258-d4d19ccc7ce7\") " pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:44 crc kubenswrapper[4848]: I1204 13:55:44.903342 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.071362 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-97c9cddfd-672td"] Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.072242 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.076150 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.076194 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.084852 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-97c9cddfd-672td"] Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.157055 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/74119917-08ea-4ef2-801d-232f7be20625-monitoring-plugin-cert\") pod \"monitoring-plugin-97c9cddfd-672td\" (UID: \"74119917-08ea-4ef2-801d-232f7be20625\") " pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.258899 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/74119917-08ea-4ef2-801d-232f7be20625-monitoring-plugin-cert\") pod \"monitoring-plugin-97c9cddfd-672td\" (UID: \"74119917-08ea-4ef2-801d-232f7be20625\") " pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.266630 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/74119917-08ea-4ef2-801d-232f7be20625-monitoring-plugin-cert\") pod \"monitoring-plugin-97c9cddfd-672td\" (UID: \"74119917-08ea-4ef2-801d-232f7be20625\") " pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.397718 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.417617 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-848dcdfb99-f4gmr"] Dec 04 13:55:45 crc kubenswrapper[4848]: W1204 13:55:45.421159 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod120107aa_e225_488c_a258_d4d19ccc7ce7.slice/crio-883c98a929bd353e6ec09dbad4e9d1968666df282739341cdf4633f3372759eb WatchSource:0}: Error finding container 883c98a929bd353e6ec09dbad4e9d1968666df282739341cdf4633f3372759eb: Status 404 returned error can't find the container with id 883c98a929bd353e6ec09dbad4e9d1968666df282739341cdf4633f3372759eb Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.641917 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.650164 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.657976 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.658113 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.658545 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-89sm5" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.661624 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.661837 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.662139 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-98titafd1heav" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.662850 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.664842 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.664889 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.664914 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.664931 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-web-config\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.664972 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665038 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhwb6\" (UniqueName: \"kubernetes.io/projected/e1d404e6-777a-483c-ab21-c750046def49-kube-api-access-qhwb6\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665079 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665103 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e1d404e6-777a-483c-ab21-c750046def49-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665154 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-config\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665171 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665185 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665200 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e1d404e6-777a-483c-ab21-c750046def49-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665514 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665546 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665567 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665636 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665727 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665773 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e1d404e6-777a-483c-ab21-c750046def49-config-out\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.665988 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-69f4c7668-cqf5s" event={"ID":"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b","Type":"ContainerStarted","Data":"3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae"} Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.666431 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-69f4c7668-cqf5s" event={"ID":"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b","Type":"ContainerStarted","Data":"e41edb87d9910c48ad45d7f3072a439b85da7c01ccd368e4da0bb6c30dcaec62"} Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.667375 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" event={"ID":"120107aa-e225-488c-a258-d4d19ccc7ce7","Type":"ContainerStarted","Data":"883c98a929bd353e6ec09dbad4e9d1968666df282739341cdf4633f3372759eb"} Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.668446 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.668609 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.668724 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.678658 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.682119 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.682912 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.683171 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.692073 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-97c9cddfd-672td"] Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767320 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767674 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e1d404e6-777a-483c-ab21-c750046def49-config-out\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767747 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767785 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767820 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767845 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767874 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-web-config\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767905 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhwb6\" (UniqueName: \"kubernetes.io/projected/e1d404e6-777a-483c-ab21-c750046def49-kube-api-access-qhwb6\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767937 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.767996 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e1d404e6-777a-483c-ab21-c750046def49-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.768026 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-config\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.768053 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.768295 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e1d404e6-777a-483c-ab21-c750046def49-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.768334 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.768375 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.768405 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.768433 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.768462 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.769247 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.776997 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e1d404e6-777a-483c-ab21-c750046def49-config-out\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.777453 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.777780 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e1d404e6-777a-483c-ab21-c750046def49-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.778036 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.779216 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.792526 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e1d404e6-777a-483c-ab21-c750046def49-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.794377 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.796606 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.796609 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-config\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.797602 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.797646 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.798002 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-web-config\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.798195 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.800604 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.801451 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e1d404e6-777a-483c-ab21-c750046def49-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.805741 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e1d404e6-777a-483c-ab21-c750046def49-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.816928 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhwb6\" (UniqueName: \"kubernetes.io/projected/e1d404e6-777a-483c-ab21-c750046def49-kube-api-access-qhwb6\") pod \"prometheus-k8s-0\" (UID: \"e1d404e6-777a-483c-ab21-c750046def49\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:45 crc kubenswrapper[4848]: I1204 13:55:45.966447 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:46 crc kubenswrapper[4848]: I1204 13:55:46.674030 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" event={"ID":"74119917-08ea-4ef2-801d-232f7be20625","Type":"ContainerStarted","Data":"35207fddd9c7c1ae3515877d841f67e92ad022136d3fd1c01d5b34668c58ce36"} Dec 04 13:55:46 crc kubenswrapper[4848]: I1204 13:55:46.696593 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-69f4c7668-cqf5s" podStartSLOduration=2.696570837 podStartE2EDuration="2.696570837s" podCreationTimestamp="2025-12-04 13:55:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:55:46.692126648 +0000 UTC m=+450.634623176" watchObservedRunningTime="2025-12-04 13:55:46.696570837 +0000 UTC m=+450.639067365" Dec 04 13:55:47 crc kubenswrapper[4848]: I1204 13:55:47.083585 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 04 13:55:47 crc kubenswrapper[4848]: W1204 13:55:47.284317 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1d404e6_777a_483c_ab21_c750046def49.slice/crio-c8e2606d609b778e99617ab64bc82afe7ed3c274f2e9ca3b842f593524d30d8f WatchSource:0}: Error finding container c8e2606d609b778e99617ab64bc82afe7ed3c274f2e9ca3b842f593524d30d8f: Status 404 returned error can't find the container with id c8e2606d609b778e99617ab64bc82afe7ed3c274f2e9ca3b842f593524d30d8f Dec 04 13:55:47 crc kubenswrapper[4848]: I1204 13:55:47.679888 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e1d404e6-777a-483c-ab21-c750046def49","Type":"ContainerStarted","Data":"c8e2606d609b778e99617ab64bc82afe7ed3c274f2e9ca3b842f593524d30d8f"} Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.687848 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8fc10b8-4f17-4ca9-affc-4578ff6fac06","Type":"ContainerStarted","Data":"1b31734acdfd454f3b6c317c505f1703264d9a76f397d4c51397389fc3f391ad"} Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.688534 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8fc10b8-4f17-4ca9-affc-4578ff6fac06","Type":"ContainerStarted","Data":"7bb0ff2b84aa7148ca005502d7ba7bfdba49063546a6aa0cf68c1a0113920bff"} Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.689629 4848 generic.go:334] "Generic (PLEG): container finished" podID="e1d404e6-777a-483c-ab21-c750046def49" containerID="48e02eacbc3a324eba14461c22f453c55b00c24054601f6ee4be448cc0349e0a" exitCode=0 Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.689688 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e1d404e6-777a-483c-ab21-c750046def49","Type":"ContainerDied","Data":"48e02eacbc3a324eba14461c22f453c55b00c24054601f6ee4be448cc0349e0a"} Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.692047 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" event={"ID":"74119917-08ea-4ef2-801d-232f7be20625","Type":"ContainerStarted","Data":"c497b2ddadf8ca63fd9baf4a0d8227c77ae9b0e2b19218f885cd5ac841b5e9a0"} Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.692189 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.695266 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" event={"ID":"568712c9-5113-4808-a56f-f58acf42eb47","Type":"ContainerStarted","Data":"934447e2bf78ca38580291fac478455df3d5c2dc81cea7f52fd3cbcbc4223115"} Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.695307 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" event={"ID":"568712c9-5113-4808-a56f-f58acf42eb47","Type":"ContainerStarted","Data":"ad2ec8584c10e5169d42e1b55a8606e3e3dab4cea3e4fff001561297ff68cb75"} Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.697101 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" event={"ID":"120107aa-e225-488c-a258-d4d19ccc7ce7","Type":"ContainerStarted","Data":"3e17628e4807a6102a8f18135743dc57a1ed960632771359966b1d42e0479213"} Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.702846 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.765484 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" podStartSLOduration=1.379383964 podStartE2EDuration="3.765455016s" podCreationTimestamp="2025-12-04 13:55:45 +0000 UTC" firstStartedPulling="2025-12-04 13:55:45.726468215 +0000 UTC m=+449.668964743" lastFinishedPulling="2025-12-04 13:55:48.112539267 +0000 UTC m=+452.055035795" observedRunningTime="2025-12-04 13:55:48.760179397 +0000 UTC m=+452.702675915" watchObservedRunningTime="2025-12-04 13:55:48.765455016 +0000 UTC m=+452.707951554" Dec 04 13:55:48 crc kubenswrapper[4848]: I1204 13:55:48.801099 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" podStartSLOduration=2.121801834 podStartE2EDuration="4.80107391s" podCreationTimestamp="2025-12-04 13:55:44 +0000 UTC" firstStartedPulling="2025-12-04 13:55:45.429730404 +0000 UTC m=+449.372226942" lastFinishedPulling="2025-12-04 13:55:48.10900249 +0000 UTC m=+452.051499018" observedRunningTime="2025-12-04 13:55:48.799580182 +0000 UTC m=+452.742076730" watchObservedRunningTime="2025-12-04 13:55:48.80107391 +0000 UTC m=+452.743570448" Dec 04 13:55:49 crc kubenswrapper[4848]: I1204 13:55:49.707850 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" event={"ID":"568712c9-5113-4808-a56f-f58acf42eb47","Type":"ContainerStarted","Data":"37f53137aff17ed3bbfa9cc15fc043633f5e42114bb6a00c68077104bef1694f"} Dec 04 13:55:49 crc kubenswrapper[4848]: I1204 13:55:49.712273 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8fc10b8-4f17-4ca9-affc-4578ff6fac06","Type":"ContainerStarted","Data":"c41c62ecf49cebd3440b8dd9e9e1918aad5b170b069211f8fde03d7f7ff1b24b"} Dec 04 13:55:49 crc kubenswrapper[4848]: I1204 13:55:49.712347 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8fc10b8-4f17-4ca9-affc-4578ff6fac06","Type":"ContainerStarted","Data":"27f558c6901c99d1adb2fffbb6791b73f3e0e70a9ce9d0ea557bcf79cd1826db"} Dec 04 13:55:49 crc kubenswrapper[4848]: I1204 13:55:49.712364 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8fc10b8-4f17-4ca9-affc-4578ff6fac06","Type":"ContainerStarted","Data":"ef38cca6dd4336b060703669db6505ed75b17483749e225f684080fe1fda2db0"} Dec 04 13:55:50 crc kubenswrapper[4848]: I1204 13:55:50.725309 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" event={"ID":"568712c9-5113-4808-a56f-f58acf42eb47","Type":"ContainerStarted","Data":"4668158567f0643234220e1815a54d3438a2c0961cc29c65d4dac955853ee115"} Dec 04 13:55:50 crc kubenswrapper[4848]: I1204 13:55:50.725558 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:55:50 crc kubenswrapper[4848]: I1204 13:55:50.725572 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" event={"ID":"568712c9-5113-4808-a56f-f58acf42eb47","Type":"ContainerStarted","Data":"0de8a5fc4c2467e0df942b69510e1711840a6dbc4eae7a1761c9366bb83ae14b"} Dec 04 13:55:50 crc kubenswrapper[4848]: I1204 13:55:50.725598 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" event={"ID":"568712c9-5113-4808-a56f-f58acf42eb47","Type":"ContainerStarted","Data":"86c74ca8047e50cb9810e7e5a8b8f2f48befc4e1f01422864a18ecb812854fb8"} Dec 04 13:55:50 crc kubenswrapper[4848]: I1204 13:55:50.731510 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8fc10b8-4f17-4ca9-affc-4578ff6fac06","Type":"ContainerStarted","Data":"0e46a5050a75f7f996e6cf4c2a210b5b80f661282e95619597ffb152ca7082c0"} Dec 04 13:55:50 crc kubenswrapper[4848]: I1204 13:55:50.753290 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" podStartSLOduration=2.182492884 podStartE2EDuration="9.753267278s" podCreationTimestamp="2025-12-04 13:55:41 +0000 UTC" firstStartedPulling="2025-12-04 13:55:42.407236706 +0000 UTC m=+446.349733234" lastFinishedPulling="2025-12-04 13:55:49.9780111 +0000 UTC m=+453.920507628" observedRunningTime="2025-12-04 13:55:50.747110977 +0000 UTC m=+454.689607555" watchObservedRunningTime="2025-12-04 13:55:50.753267278 +0000 UTC m=+454.695763806" Dec 04 13:55:50 crc kubenswrapper[4848]: I1204 13:55:50.785357 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=1.9324805619999998 podStartE2EDuration="10.785292743s" podCreationTimestamp="2025-12-04 13:55:40 +0000 UTC" firstStartedPulling="2025-12-04 13:55:41.127435284 +0000 UTC m=+445.069931812" lastFinishedPulling="2025-12-04 13:55:49.980247475 +0000 UTC m=+453.922743993" observedRunningTime="2025-12-04 13:55:50.780255609 +0000 UTC m=+454.722752157" watchObservedRunningTime="2025-12-04 13:55:50.785292743 +0000 UTC m=+454.727789291" Dec 04 13:55:52 crc kubenswrapper[4848]: I1204 13:55:52.748784 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e1d404e6-777a-483c-ab21-c750046def49","Type":"ContainerStarted","Data":"23b1efeb4af7e5cc575723d1a7103ffff8d39f40cb8fc4d82fab7ba96869379a"} Dec 04 13:55:52 crc kubenswrapper[4848]: I1204 13:55:52.749216 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e1d404e6-777a-483c-ab21-c750046def49","Type":"ContainerStarted","Data":"16e7bf77e42ec2bc5f2d39313e8bdab230962c6fc60195a89e1c74917bdafe23"} Dec 04 13:55:52 crc kubenswrapper[4848]: I1204 13:55:52.749227 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e1d404e6-777a-483c-ab21-c750046def49","Type":"ContainerStarted","Data":"743925528ffa97b52c12dedba139892a608ab39fefb2733e2cb7f25a93c092b2"} Dec 04 13:55:52 crc kubenswrapper[4848]: I1204 13:55:52.749236 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e1d404e6-777a-483c-ab21-c750046def49","Type":"ContainerStarted","Data":"9a04b114b4a5dc99ae9084286ed86feb92b59f0bd3a9ae6f139ccdaa5cec576f"} Dec 04 13:55:52 crc kubenswrapper[4848]: I1204 13:55:52.749247 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e1d404e6-777a-483c-ab21-c750046def49","Type":"ContainerStarted","Data":"9e9e570b5bce4dfc242079fe69cac5d6730ec8cf474c53d5db9731319d0f016b"} Dec 04 13:55:52 crc kubenswrapper[4848]: I1204 13:55:52.749256 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"e1d404e6-777a-483c-ab21-c750046def49","Type":"ContainerStarted","Data":"b164fd6639b652bf2dc13458cc710eb844e397d6f31e1e4cf66f7c57a70bc5ac"} Dec 04 13:55:54 crc kubenswrapper[4848]: I1204 13:55:54.411923 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:54 crc kubenswrapper[4848]: I1204 13:55:54.412126 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:54 crc kubenswrapper[4848]: I1204 13:55:54.419568 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:54 crc kubenswrapper[4848]: I1204 13:55:54.441601 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=6.341481115 podStartE2EDuration="9.441572582s" podCreationTimestamp="2025-12-04 13:55:45 +0000 UTC" firstStartedPulling="2025-12-04 13:55:48.691381551 +0000 UTC m=+452.633878079" lastFinishedPulling="2025-12-04 13:55:51.791473008 +0000 UTC m=+455.733969546" observedRunningTime="2025-12-04 13:55:52.785074671 +0000 UTC m=+456.727571239" watchObservedRunningTime="2025-12-04 13:55:54.441572582 +0000 UTC m=+458.384069120" Dec 04 13:55:54 crc kubenswrapper[4848]: I1204 13:55:54.780072 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:55:54 crc kubenswrapper[4848]: I1204 13:55:54.847537 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rrtwl"] Dec 04 13:55:55 crc kubenswrapper[4848]: I1204 13:55:55.967734 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:55:56 crc kubenswrapper[4848]: I1204 13:55:56.689747 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-779569c9f6-hdxhl" Dec 04 13:56:04 crc kubenswrapper[4848]: I1204 13:56:04.903900 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:56:04 crc kubenswrapper[4848]: I1204 13:56:04.904544 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:56:19 crc kubenswrapper[4848]: I1204 13:56:19.897064 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-rrtwl" podUID="d175633c-17b1-48b5-a1ed-430f08118b6c" containerName="console" containerID="cri-o://70c0021c264792e5830b8afe4b00288702b71b7e8492b397fbe99ea80c68e1ef" gracePeriod=15 Dec 04 13:56:21 crc kubenswrapper[4848]: I1204 13:56:21.964996 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rrtwl_d175633c-17b1-48b5-a1ed-430f08118b6c/console/0.log" Dec 04 13:56:21 crc kubenswrapper[4848]: I1204 13:56:21.965397 4848 generic.go:334] "Generic (PLEG): container finished" podID="d175633c-17b1-48b5-a1ed-430f08118b6c" containerID="70c0021c264792e5830b8afe4b00288702b71b7e8492b397fbe99ea80c68e1ef" exitCode=2 Dec 04 13:56:21 crc kubenswrapper[4848]: I1204 13:56:21.965448 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rrtwl" event={"ID":"d175633c-17b1-48b5-a1ed-430f08118b6c","Type":"ContainerDied","Data":"70c0021c264792e5830b8afe4b00288702b71b7e8492b397fbe99ea80c68e1ef"} Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.412212 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rrtwl_d175633c-17b1-48b5-a1ed-430f08118b6c/console/0.log" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.412505 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.551698 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-trusted-ca-bundle\") pod \"d175633c-17b1-48b5-a1ed-430f08118b6c\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.551761 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-console-config\") pod \"d175633c-17b1-48b5-a1ed-430f08118b6c\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.551831 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-oauth-serving-cert\") pod \"d175633c-17b1-48b5-a1ed-430f08118b6c\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.552011 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-oauth-config\") pod \"d175633c-17b1-48b5-a1ed-430f08118b6c\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.552079 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-service-ca\") pod \"d175633c-17b1-48b5-a1ed-430f08118b6c\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.552116 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjm4m\" (UniqueName: \"kubernetes.io/projected/d175633c-17b1-48b5-a1ed-430f08118b6c-kube-api-access-vjm4m\") pod \"d175633c-17b1-48b5-a1ed-430f08118b6c\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.552150 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-serving-cert\") pod \"d175633c-17b1-48b5-a1ed-430f08118b6c\" (UID: \"d175633c-17b1-48b5-a1ed-430f08118b6c\") " Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.553457 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-service-ca" (OuterVolumeSpecName: "service-ca") pod "d175633c-17b1-48b5-a1ed-430f08118b6c" (UID: "d175633c-17b1-48b5-a1ed-430f08118b6c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.553482 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d175633c-17b1-48b5-a1ed-430f08118b6c" (UID: "d175633c-17b1-48b5-a1ed-430f08118b6c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.553491 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d175633c-17b1-48b5-a1ed-430f08118b6c" (UID: "d175633c-17b1-48b5-a1ed-430f08118b6c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.553696 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-console-config" (OuterVolumeSpecName: "console-config") pod "d175633c-17b1-48b5-a1ed-430f08118b6c" (UID: "d175633c-17b1-48b5-a1ed-430f08118b6c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.559180 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d175633c-17b1-48b5-a1ed-430f08118b6c" (UID: "d175633c-17b1-48b5-a1ed-430f08118b6c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.561202 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d175633c-17b1-48b5-a1ed-430f08118b6c" (UID: "d175633c-17b1-48b5-a1ed-430f08118b6c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.561866 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d175633c-17b1-48b5-a1ed-430f08118b6c-kube-api-access-vjm4m" (OuterVolumeSpecName: "kube-api-access-vjm4m") pod "d175633c-17b1-48b5-a1ed-430f08118b6c" (UID: "d175633c-17b1-48b5-a1ed-430f08118b6c"). InnerVolumeSpecName "kube-api-access-vjm4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.654628 4848 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.654697 4848 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.654724 4848 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.654745 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjm4m\" (UniqueName: \"kubernetes.io/projected/d175633c-17b1-48b5-a1ed-430f08118b6c-kube-api-access-vjm4m\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.654768 4848 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d175633c-17b1-48b5-a1ed-430f08118b6c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.654789 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.654810 4848 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d175633c-17b1-48b5-a1ed-430f08118b6c-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.975215 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rrtwl_d175633c-17b1-48b5-a1ed-430f08118b6c/console/0.log" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.975280 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rrtwl" event={"ID":"d175633c-17b1-48b5-a1ed-430f08118b6c","Type":"ContainerDied","Data":"aa5b70434d977b36b82fbdb39edeb8b21eddf525ffb7a790454b8618ee94319b"} Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.975354 4848 scope.go:117] "RemoveContainer" containerID="70c0021c264792e5830b8afe4b00288702b71b7e8492b397fbe99ea80c68e1ef" Dec 04 13:56:22 crc kubenswrapper[4848]: I1204 13:56:22.975369 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rrtwl" Dec 04 13:56:23 crc kubenswrapper[4848]: I1204 13:56:23.011802 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rrtwl"] Dec 04 13:56:23 crc kubenswrapper[4848]: I1204 13:56:23.018592 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-rrtwl"] Dec 04 13:56:24 crc kubenswrapper[4848]: I1204 13:56:24.410588 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d175633c-17b1-48b5-a1ed-430f08118b6c" path="/var/lib/kubelet/pods/d175633c-17b1-48b5-a1ed-430f08118b6c/volumes" Dec 04 13:56:24 crc kubenswrapper[4848]: I1204 13:56:24.911295 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:56:24 crc kubenswrapper[4848]: I1204 13:56:24.916248 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-848dcdfb99-f4gmr" Dec 04 13:56:45 crc kubenswrapper[4848]: I1204 13:56:45.967380 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:56:46 crc kubenswrapper[4848]: I1204 13:56:46.005283 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:56:46 crc kubenswrapper[4848]: I1204 13:56:46.193325 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.711156 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-74cc9ddcc4-dx5vs"] Dec 04 13:56:57 crc kubenswrapper[4848]: E1204 13:56:57.711810 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d175633c-17b1-48b5-a1ed-430f08118b6c" containerName="console" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.711821 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d175633c-17b1-48b5-a1ed-430f08118b6c" containerName="console" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.711935 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d175633c-17b1-48b5-a1ed-430f08118b6c" containerName="console" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.712351 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.726434 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74cc9ddcc4-dx5vs"] Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.726888 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-oauth-serving-cert\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.726989 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-serving-cert\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.727099 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-config\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.727142 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jg2v\" (UniqueName: \"kubernetes.io/projected/9227885e-0ea7-47fa-a763-3a3dd3c1de39-kube-api-access-7jg2v\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.727198 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-service-ca\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.727246 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-trusted-ca-bundle\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.727287 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-oauth-config\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.828477 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-config\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.828706 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jg2v\" (UniqueName: \"kubernetes.io/projected/9227885e-0ea7-47fa-a763-3a3dd3c1de39-kube-api-access-7jg2v\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.828800 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-service-ca\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.828878 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-trusted-ca-bundle\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.828965 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-oauth-config\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.829084 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-oauth-serving-cert\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.829160 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-serving-cert\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.829668 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-config\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.829886 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-service-ca\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.830466 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-trusted-ca-bundle\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.830475 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-oauth-serving-cert\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.833817 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-oauth-config\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.833905 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-serving-cert\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:57 crc kubenswrapper[4848]: I1204 13:56:57.852458 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jg2v\" (UniqueName: \"kubernetes.io/projected/9227885e-0ea7-47fa-a763-3a3dd3c1de39-kube-api-access-7jg2v\") pod \"console-74cc9ddcc4-dx5vs\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:58 crc kubenswrapper[4848]: I1204 13:56:58.034883 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:56:58 crc kubenswrapper[4848]: I1204 13:56:58.498975 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74cc9ddcc4-dx5vs"] Dec 04 13:56:59 crc kubenswrapper[4848]: I1204 13:56:59.233093 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74cc9ddcc4-dx5vs" event={"ID":"9227885e-0ea7-47fa-a763-3a3dd3c1de39","Type":"ContainerStarted","Data":"4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094"} Dec 04 13:56:59 crc kubenswrapper[4848]: I1204 13:56:59.233128 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74cc9ddcc4-dx5vs" event={"ID":"9227885e-0ea7-47fa-a763-3a3dd3c1de39","Type":"ContainerStarted","Data":"d5f63630cda22165cfe0c5d5846d9bbe3f7c614af160ec66c8d41bb20e7daadf"} Dec 04 13:56:59 crc kubenswrapper[4848]: I1204 13:56:59.256513 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-74cc9ddcc4-dx5vs" podStartSLOduration=2.256491704 podStartE2EDuration="2.256491704s" podCreationTimestamp="2025-12-04 13:56:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:56:59.252000214 +0000 UTC m=+523.194496762" watchObservedRunningTime="2025-12-04 13:56:59.256491704 +0000 UTC m=+523.198988252" Dec 04 13:57:08 crc kubenswrapper[4848]: I1204 13:57:08.035871 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:57:08 crc kubenswrapper[4848]: I1204 13:57:08.036666 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:57:08 crc kubenswrapper[4848]: I1204 13:57:08.040905 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:57:08 crc kubenswrapper[4848]: I1204 13:57:08.303433 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 13:57:08 crc kubenswrapper[4848]: I1204 13:57:08.384694 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-69f4c7668-cqf5s"] Dec 04 13:57:16 crc kubenswrapper[4848]: I1204 13:57:16.573496 4848 scope.go:117] "RemoveContainer" containerID="c5b4cb4faaea6a1fa269389d85f5f2cbeb9d4d4fd0a8d7aa390102ca8042950c" Dec 04 13:57:33 crc kubenswrapper[4848]: I1204 13:57:33.440684 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-69f4c7668-cqf5s" podUID="02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" containerName="console" containerID="cri-o://3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae" gracePeriod=15 Dec 04 13:57:33 crc kubenswrapper[4848]: I1204 13:57:33.876996 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-69f4c7668-cqf5s_02eac971-3e5f-4d8b-baab-5a1e68f0ac7b/console/0.log" Dec 04 13:57:33 crc kubenswrapper[4848]: I1204 13:57:33.877321 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.023653 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-config\") pod \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.023854 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-oauth-serving-cert\") pod \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.024622 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-trusted-ca-bundle\") pod \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.024655 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-config" (OuterVolumeSpecName: "console-config") pod "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" (UID: "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.024737 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-serving-cert\") pod \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.024796 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnqwc\" (UniqueName: \"kubernetes.io/projected/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-kube-api-access-dnqwc\") pod \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.024830 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" (UID: "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.024848 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-oauth-config\") pod \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.025207 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" (UID: "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.025253 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-service-ca\") pod \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\" (UID: \"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b\") " Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.025576 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-service-ca" (OuterVolumeSpecName: "service-ca") pod "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" (UID: "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.025767 4848 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.025806 4848 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.025828 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.025848 4848 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.030344 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-kube-api-access-dnqwc" (OuterVolumeSpecName: "kube-api-access-dnqwc") pod "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" (UID: "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b"). InnerVolumeSpecName "kube-api-access-dnqwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.030679 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" (UID: "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.031539 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" (UID: "02eac971-3e5f-4d8b-baab-5a1e68f0ac7b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.127047 4848 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.127119 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnqwc\" (UniqueName: \"kubernetes.io/projected/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-kube-api-access-dnqwc\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.127140 4848 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.560666 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-69f4c7668-cqf5s_02eac971-3e5f-4d8b-baab-5a1e68f0ac7b/console/0.log" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.560767 4848 generic.go:334] "Generic (PLEG): container finished" podID="02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" containerID="3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae" exitCode=2 Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.560814 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-69f4c7668-cqf5s" event={"ID":"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b","Type":"ContainerDied","Data":"3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae"} Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.560843 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-69f4c7668-cqf5s" event={"ID":"02eac971-3e5f-4d8b-baab-5a1e68f0ac7b","Type":"ContainerDied","Data":"e41edb87d9910c48ad45d7f3072a439b85da7c01ccd368e4da0bb6c30dcaec62"} Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.560860 4848 scope.go:117] "RemoveContainer" containerID="3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.560903 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-69f4c7668-cqf5s" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.597282 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-69f4c7668-cqf5s"] Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.601145 4848 scope.go:117] "RemoveContainer" containerID="3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae" Dec 04 13:57:34 crc kubenswrapper[4848]: E1204 13:57:34.601560 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae\": container with ID starting with 3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae not found: ID does not exist" containerID="3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.601592 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae"} err="failed to get container status \"3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae\": rpc error: code = NotFound desc = could not find container \"3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae\": container with ID starting with 3ccaa0f86289ac7d850c39fe762e1efcdee6692bbb11b9dea1aeffc2239066ae not found: ID does not exist" Dec 04 13:57:34 crc kubenswrapper[4848]: I1204 13:57:34.607516 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-69f4c7668-cqf5s"] Dec 04 13:57:36 crc kubenswrapper[4848]: I1204 13:57:36.404313 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" path="/var/lib/kubelet/pods/02eac971-3e5f-4d8b-baab-5a1e68f0ac7b/volumes" Dec 04 13:57:44 crc kubenswrapper[4848]: I1204 13:57:44.315146 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:57:44 crc kubenswrapper[4848]: I1204 13:57:44.315737 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:58:14 crc kubenswrapper[4848]: I1204 13:58:14.314331 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:58:14 crc kubenswrapper[4848]: I1204 13:58:14.315363 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:58:16 crc kubenswrapper[4848]: I1204 13:58:16.647264 4848 scope.go:117] "RemoveContainer" containerID="0a090d37c3da4dbf63f6d6e5ade249782fd8254f8e48cd6c92e463c134cc8324" Dec 04 13:58:16 crc kubenswrapper[4848]: I1204 13:58:16.678828 4848 scope.go:117] "RemoveContainer" containerID="618cdd637d5200c2e36b73fee806765f20debe62604e7e3b00597ee7513209e9" Dec 04 13:58:44 crc kubenswrapper[4848]: I1204 13:58:44.315366 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:58:44 crc kubenswrapper[4848]: I1204 13:58:44.315983 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:58:44 crc kubenswrapper[4848]: I1204 13:58:44.316050 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 13:58:44 crc kubenswrapper[4848]: I1204 13:58:44.316930 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c3cebfe15e9ee005602e42c9fe3958ff03b1e9bc1b525474da30cb9fd280b5af"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:58:44 crc kubenswrapper[4848]: I1204 13:58:44.317072 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://c3cebfe15e9ee005602e42c9fe3958ff03b1e9bc1b525474da30cb9fd280b5af" gracePeriod=600 Dec 04 13:58:45 crc kubenswrapper[4848]: I1204 13:58:45.049555 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="c3cebfe15e9ee005602e42c9fe3958ff03b1e9bc1b525474da30cb9fd280b5af" exitCode=0 Dec 04 13:58:45 crc kubenswrapper[4848]: I1204 13:58:45.049732 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"c3cebfe15e9ee005602e42c9fe3958ff03b1e9bc1b525474da30cb9fd280b5af"} Dec 04 13:58:45 crc kubenswrapper[4848]: I1204 13:58:45.050346 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"97feeafde380711ec8873168a8a64ca37c03edc0904564e6ca0465fe48907206"} Dec 04 13:58:45 crc kubenswrapper[4848]: I1204 13:58:45.050382 4848 scope.go:117] "RemoveContainer" containerID="72d765596d3c5a2d39e725591250e92b81042c983d061e6907f508172e90aa2a" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.165398 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps"] Dec 04 14:00:00 crc kubenswrapper[4848]: E1204 14:00:00.167711 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" containerName="console" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.167795 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" containerName="console" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.167989 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="02eac971-3e5f-4d8b-baab-5a1e68f0ac7b" containerName="console" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.168844 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.170472 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.174315 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps"] Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.214123 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.314886 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790c562-9bc1-4b46-bea9-817210a2658a-secret-volume\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.315202 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv9m6\" (UniqueName: \"kubernetes.io/projected/e790c562-9bc1-4b46-bea9-817210a2658a-kube-api-access-sv9m6\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.315272 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790c562-9bc1-4b46-bea9-817210a2658a-config-volume\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.416908 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790c562-9bc1-4b46-bea9-817210a2658a-secret-volume\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.417098 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv9m6\" (UniqueName: \"kubernetes.io/projected/e790c562-9bc1-4b46-bea9-817210a2658a-kube-api-access-sv9m6\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.417148 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790c562-9bc1-4b46-bea9-817210a2658a-config-volume\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.418782 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790c562-9bc1-4b46-bea9-817210a2658a-config-volume\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.429311 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790c562-9bc1-4b46-bea9-817210a2658a-secret-volume\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.441224 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv9m6\" (UniqueName: \"kubernetes.io/projected/e790c562-9bc1-4b46-bea9-817210a2658a-kube-api-access-sv9m6\") pod \"collect-profiles-29414280-dgwps\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.529227 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:00 crc kubenswrapper[4848]: I1204 14:00:00.723597 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps"] Dec 04 14:00:01 crc kubenswrapper[4848]: I1204 14:00:01.624293 4848 generic.go:334] "Generic (PLEG): container finished" podID="e790c562-9bc1-4b46-bea9-817210a2658a" containerID="1ef62e11ff1abc4703fc14131b12f21368c4cd30c0c3878948c0446505378348" exitCode=0 Dec 04 14:00:01 crc kubenswrapper[4848]: I1204 14:00:01.624344 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" event={"ID":"e790c562-9bc1-4b46-bea9-817210a2658a","Type":"ContainerDied","Data":"1ef62e11ff1abc4703fc14131b12f21368c4cd30c0c3878948c0446505378348"} Dec 04 14:00:01 crc kubenswrapper[4848]: I1204 14:00:01.624377 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" event={"ID":"e790c562-9bc1-4b46-bea9-817210a2658a","Type":"ContainerStarted","Data":"8dee048a902340dcc60b646fc98f23e34428734e5804540e9ab67a12613434f5"} Dec 04 14:00:02 crc kubenswrapper[4848]: I1204 14:00:02.831011 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:02 crc kubenswrapper[4848]: I1204 14:00:02.951252 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv9m6\" (UniqueName: \"kubernetes.io/projected/e790c562-9bc1-4b46-bea9-817210a2658a-kube-api-access-sv9m6\") pod \"e790c562-9bc1-4b46-bea9-817210a2658a\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " Dec 04 14:00:02 crc kubenswrapper[4848]: I1204 14:00:02.951331 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790c562-9bc1-4b46-bea9-817210a2658a-config-volume\") pod \"e790c562-9bc1-4b46-bea9-817210a2658a\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " Dec 04 14:00:02 crc kubenswrapper[4848]: I1204 14:00:02.951377 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790c562-9bc1-4b46-bea9-817210a2658a-secret-volume\") pod \"e790c562-9bc1-4b46-bea9-817210a2658a\" (UID: \"e790c562-9bc1-4b46-bea9-817210a2658a\") " Dec 04 14:00:02 crc kubenswrapper[4848]: I1204 14:00:02.952157 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e790c562-9bc1-4b46-bea9-817210a2658a-config-volume" (OuterVolumeSpecName: "config-volume") pod "e790c562-9bc1-4b46-bea9-817210a2658a" (UID: "e790c562-9bc1-4b46-bea9-817210a2658a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:00:02 crc kubenswrapper[4848]: I1204 14:00:02.956337 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e790c562-9bc1-4b46-bea9-817210a2658a-kube-api-access-sv9m6" (OuterVolumeSpecName: "kube-api-access-sv9m6") pod "e790c562-9bc1-4b46-bea9-817210a2658a" (UID: "e790c562-9bc1-4b46-bea9-817210a2658a"). InnerVolumeSpecName "kube-api-access-sv9m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:00:02 crc kubenswrapper[4848]: I1204 14:00:02.960359 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e790c562-9bc1-4b46-bea9-817210a2658a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e790c562-9bc1-4b46-bea9-817210a2658a" (UID: "e790c562-9bc1-4b46-bea9-817210a2658a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:00:03 crc kubenswrapper[4848]: I1204 14:00:03.052405 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790c562-9bc1-4b46-bea9-817210a2658a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:03 crc kubenswrapper[4848]: I1204 14:00:03.052443 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790c562-9bc1-4b46-bea9-817210a2658a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:03 crc kubenswrapper[4848]: I1204 14:00:03.052452 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv9m6\" (UniqueName: \"kubernetes.io/projected/e790c562-9bc1-4b46-bea9-817210a2658a-kube-api-access-sv9m6\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:03 crc kubenswrapper[4848]: I1204 14:00:03.639969 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" event={"ID":"e790c562-9bc1-4b46-bea9-817210a2658a","Type":"ContainerDied","Data":"8dee048a902340dcc60b646fc98f23e34428734e5804540e9ab67a12613434f5"} Dec 04 14:00:03 crc kubenswrapper[4848]: I1204 14:00:03.640263 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dee048a902340dcc60b646fc98f23e34428734e5804540e9ab67a12613434f5" Dec 04 14:00:03 crc kubenswrapper[4848]: I1204 14:00:03.640072 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.486509 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl"] Dec 04 14:00:41 crc kubenswrapper[4848]: E1204 14:00:41.488276 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e790c562-9bc1-4b46-bea9-817210a2658a" containerName="collect-profiles" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.488364 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="e790c562-9bc1-4b46-bea9-817210a2658a" containerName="collect-profiles" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.488553 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="e790c562-9bc1-4b46-bea9-817210a2658a" containerName="collect-profiles" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.489559 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.491391 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.496570 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl"] Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.610651 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.610724 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78f75\" (UniqueName: \"kubernetes.io/projected/f76a919c-d711-434c-847f-20a25cecfd57-kube-api-access-78f75\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.610875 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.712269 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.712393 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.712415 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78f75\" (UniqueName: \"kubernetes.io/projected/f76a919c-d711-434c-847f-20a25cecfd57-kube-api-access-78f75\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.712864 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.713006 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.739203 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78f75\" (UniqueName: \"kubernetes.io/projected/f76a919c-d711-434c-847f-20a25cecfd57-kube-api-access-78f75\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:41 crc kubenswrapper[4848]: I1204 14:00:41.810427 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:42 crc kubenswrapper[4848]: I1204 14:00:42.073122 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl"] Dec 04 14:00:42 crc kubenswrapper[4848]: I1204 14:00:42.904058 4848 generic.go:334] "Generic (PLEG): container finished" podID="f76a919c-d711-434c-847f-20a25cecfd57" containerID="20a78177c9395269b94f4874904ff5fd584ff3136cc70876853c489e26651763" exitCode=0 Dec 04 14:00:42 crc kubenswrapper[4848]: I1204 14:00:42.904106 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" event={"ID":"f76a919c-d711-434c-847f-20a25cecfd57","Type":"ContainerDied","Data":"20a78177c9395269b94f4874904ff5fd584ff3136cc70876853c489e26651763"} Dec 04 14:00:42 crc kubenswrapper[4848]: I1204 14:00:42.904135 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" event={"ID":"f76a919c-d711-434c-847f-20a25cecfd57","Type":"ContainerStarted","Data":"51bdb0c3c3fbf1f0f697ae318c12654d64b71d17ef5da2aff70734ab1ad6331c"} Dec 04 14:00:42 crc kubenswrapper[4848]: I1204 14:00:42.906538 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:00:44 crc kubenswrapper[4848]: I1204 14:00:44.314246 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:00:44 crc kubenswrapper[4848]: I1204 14:00:44.315004 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:00:44 crc kubenswrapper[4848]: I1204 14:00:44.922246 4848 generic.go:334] "Generic (PLEG): container finished" podID="f76a919c-d711-434c-847f-20a25cecfd57" containerID="dfd4a5b803bfea3ecdf52157a2614708aef39b501bb26fccaf57454a9fb72038" exitCode=0 Dec 04 14:00:44 crc kubenswrapper[4848]: I1204 14:00:44.922302 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" event={"ID":"f76a919c-d711-434c-847f-20a25cecfd57","Type":"ContainerDied","Data":"dfd4a5b803bfea3ecdf52157a2614708aef39b501bb26fccaf57454a9fb72038"} Dec 04 14:00:45 crc kubenswrapper[4848]: I1204 14:00:45.930238 4848 generic.go:334] "Generic (PLEG): container finished" podID="f76a919c-d711-434c-847f-20a25cecfd57" containerID="2b26944852a112cfceb5853dd7ad705da8cc05173b4524265847e33ee80136d7" exitCode=0 Dec 04 14:00:45 crc kubenswrapper[4848]: I1204 14:00:45.930313 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" event={"ID":"f76a919c-d711-434c-847f-20a25cecfd57","Type":"ContainerDied","Data":"2b26944852a112cfceb5853dd7ad705da8cc05173b4524265847e33ee80136d7"} Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.194624 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.211060 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-util\") pod \"f76a919c-d711-434c-847f-20a25cecfd57\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.211140 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-bundle\") pod \"f76a919c-d711-434c-847f-20a25cecfd57\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.211185 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78f75\" (UniqueName: \"kubernetes.io/projected/f76a919c-d711-434c-847f-20a25cecfd57-kube-api-access-78f75\") pod \"f76a919c-d711-434c-847f-20a25cecfd57\" (UID: \"f76a919c-d711-434c-847f-20a25cecfd57\") " Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.213549 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-bundle" (OuterVolumeSpecName: "bundle") pod "f76a919c-d711-434c-847f-20a25cecfd57" (UID: "f76a919c-d711-434c-847f-20a25cecfd57"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.231421 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-util" (OuterVolumeSpecName: "util") pod "f76a919c-d711-434c-847f-20a25cecfd57" (UID: "f76a919c-d711-434c-847f-20a25cecfd57"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.235495 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f76a919c-d711-434c-847f-20a25cecfd57-kube-api-access-78f75" (OuterVolumeSpecName: "kube-api-access-78f75") pod "f76a919c-d711-434c-847f-20a25cecfd57" (UID: "f76a919c-d711-434c-847f-20a25cecfd57"). InnerVolumeSpecName "kube-api-access-78f75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.313001 4848 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.313046 4848 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f76a919c-d711-434c-847f-20a25cecfd57-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.313057 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78f75\" (UniqueName: \"kubernetes.io/projected/f76a919c-d711-434c-847f-20a25cecfd57-kube-api-access-78f75\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.945301 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" event={"ID":"f76a919c-d711-434c-847f-20a25cecfd57","Type":"ContainerDied","Data":"51bdb0c3c3fbf1f0f697ae318c12654d64b71d17ef5da2aff70734ab1ad6331c"} Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.945542 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51bdb0c3c3fbf1f0f697ae318c12654d64b71d17ef5da2aff70734ab1ad6331c" Dec 04 14:00:47 crc kubenswrapper[4848]: I1204 14:00:47.945360 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl" Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.473648 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vwdkp"] Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.474624 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovn-controller" containerID="cri-o://9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5" gracePeriod=30 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.474687 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="nbdb" containerID="cri-o://17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" gracePeriod=30 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.474747 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovn-acl-logging" containerID="cri-o://a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9" gracePeriod=30 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.475128 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db" gracePeriod=30 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.475202 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kube-rbac-proxy-node" containerID="cri-o://4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390" gracePeriod=30 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.475344 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="sbdb" containerID="cri-o://ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" gracePeriod=30 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.475407 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="northd" containerID="cri-o://89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591" gracePeriod=30 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.519658 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" containerID="cri-o://21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239" gracePeriod=30 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.979795 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovnkube-controller/3.log" Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.983112 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovn-acl-logging/0.log" Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.985378 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovn-controller/0.log" Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.986185 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239" exitCode=0 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.986428 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" exitCode=0 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.986611 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" exitCode=0 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.986771 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591" exitCode=0 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.986930 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9" exitCode=143 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.987120 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5" exitCode=143 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.986286 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239"} Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.987575 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6"} Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.987781 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907"} Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.988002 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591"} Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.988178 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9"} Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.988355 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5"} Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.987636 4848 scope.go:117] "RemoveContainer" containerID="83e63512da8aded52282ffafdfa55dca3f2c1838241d79cde0190dfb3de8aa02" Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.990453 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/2.log" Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.992193 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/1.log" Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.992283 4848 generic.go:334] "Generic (PLEG): container finished" podID="2b6bff84-ab72-4936-8c9f-0508ded8fdd8" containerID="353f581d9b32e9f053062e13bc3dc93e3a02c4028418b5a55aba4ce1fb70fc53" exitCode=2 Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.992328 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spvsd" event={"ID":"2b6bff84-ab72-4936-8c9f-0508ded8fdd8","Type":"ContainerDied","Data":"353f581d9b32e9f053062e13bc3dc93e3a02c4028418b5a55aba4ce1fb70fc53"} Dec 04 14:00:52 crc kubenswrapper[4848]: I1204 14:00:52.993147 4848 scope.go:117] "RemoveContainer" containerID="353f581d9b32e9f053062e13bc3dc93e3a02c4028418b5a55aba4ce1fb70fc53" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.161237 4848 scope.go:117] "RemoveContainer" containerID="5659fe913de27f42cf6336ab5f3cb58484604e65c740f833af2fec12ea7af955" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.461626 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907 is running failed: container process not found" containerID="17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.462249 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6 is running failed: container process not found" containerID="ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.462550 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907 is running failed: container process not found" containerID="17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.462643 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6 is running failed: container process not found" containerID="ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.462998 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907 is running failed: container process not found" containerID="17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.463042 4848 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="nbdb" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.463061 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6 is running failed: container process not found" containerID="ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.463093 4848 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="sbdb" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.711934 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovn-acl-logging/0.log" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.713505 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovn-controller/0.log" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.713962 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794143 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vjh9w"] Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794359 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794370 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794379 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kubecfg-setup" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794385 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kubecfg-setup" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794396 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="nbdb" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794403 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="nbdb" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794413 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794419 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794428 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f76a919c-d711-434c-847f-20a25cecfd57" containerName="extract" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794434 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f76a919c-d711-434c-847f-20a25cecfd57" containerName="extract" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794445 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794450 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794457 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f76a919c-d711-434c-847f-20a25cecfd57" containerName="util" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794463 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f76a919c-d711-434c-847f-20a25cecfd57" containerName="util" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794469 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794475 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794483 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kube-rbac-proxy-node" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794488 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kube-rbac-proxy-node" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794500 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovn-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794505 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovn-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794511 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovn-acl-logging" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794517 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovn-acl-logging" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794524 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="sbdb" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794529 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="sbdb" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794538 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="northd" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794543 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="northd" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794551 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794557 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794564 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f76a919c-d711-434c-847f-20a25cecfd57" containerName="pull" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794571 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f76a919c-d711-434c-847f-20a25cecfd57" containerName="pull" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794673 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794685 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794692 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794698 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="northd" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794704 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovn-acl-logging" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794712 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="sbdb" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794719 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794727 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794736 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovn-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794745 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="nbdb" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794752 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="kube-rbac-proxy-node" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794759 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f76a919c-d711-434c-847f-20a25cecfd57" containerName="extract" Dec 04 14:00:53 crc kubenswrapper[4848]: E1204 14:00:53.794854 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794862 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.794974 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88de7a37-3579-4630-9aa7-64654e68a472" containerName="ovnkube-controller" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.796640 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.895960 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-script-lib\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896018 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-config\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896039 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-openvswitch\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896062 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88de7a37-3579-4630-9aa7-64654e68a472-ovn-node-metrics-cert\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896086 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-ovn\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896139 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896141 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896210 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-systemd\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896484 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896546 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-log-socket\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896576 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-ovn-kubernetes\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896591 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-etc-openvswitch\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896611 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-bin\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896625 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-slash\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896642 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-var-lib-cni-networks-ovn-kubernetes\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896655 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-log-socket" (OuterVolumeSpecName: "log-socket") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896664 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-netns\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896681 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896685 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-netd\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896682 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896703 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896710 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-systemd-units\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896725 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-kubelet\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896727 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896752 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-env-overrides\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896731 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896742 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896770 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-node-log\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896772 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896813 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896815 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-node-log" (OuterVolumeSpecName: "node-log") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896794 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-var-lib-openvswitch\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896788 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-slash" (OuterVolumeSpecName: "host-slash") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896775 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.896927 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7z6f\" (UniqueName: \"kubernetes.io/projected/88de7a37-3579-4630-9aa7-64654e68a472-kube-api-access-h7z6f\") pod \"88de7a37-3579-4630-9aa7-64654e68a472\" (UID: \"88de7a37-3579-4630-9aa7-64654e68a472\") " Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897006 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897151 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897247 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-cni-netd\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897285 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-node-log\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897313 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-log-socket\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897367 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mqs7\" (UniqueName: \"kubernetes.io/projected/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-kube-api-access-2mqs7\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897398 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-kubelet\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897461 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-systemd\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897482 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897535 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897570 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-cni-bin\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897612 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovnkube-script-lib\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897635 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-slash\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897699 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-env-overrides\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897741 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-etc-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897768 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovnkube-config\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897804 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-ovn\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897845 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-run-ovn-kubernetes\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897874 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-systemd-units\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897901 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-var-lib-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897931 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovn-node-metrics-cert\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.897972 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-run-netns\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898020 4848 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898033 4848 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898045 4848 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898055 4848 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-node-log\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898064 4848 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898073 4848 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898082 4848 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88de7a37-3579-4630-9aa7-64654e68a472-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898090 4848 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898099 4848 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898107 4848 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-log-socket\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898115 4848 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898127 4848 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898135 4848 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898142 4848 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-slash\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898153 4848 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898164 4848 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.898172 4848 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.906955 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88de7a37-3579-4630-9aa7-64654e68a472-kube-api-access-h7z6f" (OuterVolumeSpecName: "kube-api-access-h7z6f") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "kube-api-access-h7z6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.910193 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88de7a37-3579-4630-9aa7-64654e68a472-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.928550 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "88de7a37-3579-4630-9aa7-64654e68a472" (UID: "88de7a37-3579-4630-9aa7-64654e68a472"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998516 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-kubelet\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998576 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-systemd\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998597 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998631 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998658 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-cni-bin\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998683 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovnkube-script-lib\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998680 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-kubelet\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998704 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-slash\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998735 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998745 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-slash\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998749 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998784 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-cni-bin\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998816 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-env-overrides\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998880 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-etc-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998919 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovnkube-config\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998963 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-ovn\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.998972 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-etc-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999002 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-ovn\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999016 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-run-ovn-kubernetes\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999042 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-systemd-units\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999065 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-var-lib-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999066 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-run-ovn-kubernetes\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999072 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-run-systemd\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999105 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-var-lib-openvswitch\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999113 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovn-node-metrics-cert\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999141 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-run-netns\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999146 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-systemd-units\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999160 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-cni-netd\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999181 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-cni-netd\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999202 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-host-run-netns\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999283 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-node-log\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999314 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-log-socket\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999337 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mqs7\" (UniqueName: \"kubernetes.io/projected/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-kube-api-access-2mqs7\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999382 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7z6f\" (UniqueName: \"kubernetes.io/projected/88de7a37-3579-4630-9aa7-64654e68a472-kube-api-access-h7z6f\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999395 4848 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88de7a37-3579-4630-9aa7-64654e68a472-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999407 4848 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/88de7a37-3579-4630-9aa7-64654e68a472-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:53 crc kubenswrapper[4848]: I1204 14:00:53.999471 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-env-overrides\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:53.999520 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-node-log\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:53.999549 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-log-socket\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:53.999573 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovnkube-script-lib\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:53.999592 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovnkube-config\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.004345 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-ovn-node-metrics-cert\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.005228 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovn-acl-logging/0.log" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.005690 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-vwdkp_88de7a37-3579-4630-9aa7-64654e68a472/ovn-controller/0.log" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.006077 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db" exitCode=0 Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.006104 4848 generic.go:334] "Generic (PLEG): container finished" podID="88de7a37-3579-4630-9aa7-64654e68a472" containerID="4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390" exitCode=0 Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.006166 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.006176 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db"} Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.006212 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390"} Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.006225 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwdkp" event={"ID":"88de7a37-3579-4630-9aa7-64654e68a472","Type":"ContainerDied","Data":"832c55b9c6aa9f5b5af6c6e9037974a5535e6a520ab26377463c2ee32070d0cd"} Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.006231 4848 scope.go:117] "RemoveContainer" containerID="21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.013471 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spvsd_2b6bff84-ab72-4936-8c9f-0508ded8fdd8/kube-multus/2.log" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.013532 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spvsd" event={"ID":"2b6bff84-ab72-4936-8c9f-0508ded8fdd8","Type":"ContainerStarted","Data":"4d1a23caa80100feb235875569f36475a4ab95352af61e540899c2196ed2ee34"} Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.028454 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mqs7\" (UniqueName: \"kubernetes.io/projected/eadb015a-0f6b-451c-a9c4-f3ccd6a6927c-kube-api-access-2mqs7\") pod \"ovnkube-node-vjh9w\" (UID: \"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c\") " pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.031031 4848 scope.go:117] "RemoveContainer" containerID="ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.046914 4848 scope.go:117] "RemoveContainer" containerID="17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.068717 4848 scope.go:117] "RemoveContainer" containerID="89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.070042 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vwdkp"] Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.078331 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vwdkp"] Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.083535 4848 scope.go:117] "RemoveContainer" containerID="5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.098843 4848 scope.go:117] "RemoveContainer" containerID="4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.110426 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.111640 4848 scope.go:117] "RemoveContainer" containerID="a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.144431 4848 scope.go:117] "RemoveContainer" containerID="9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.163314 4848 scope.go:117] "RemoveContainer" containerID="f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.182799 4848 scope.go:117] "RemoveContainer" containerID="21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.183313 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239\": container with ID starting with 21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239 not found: ID does not exist" containerID="21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.183356 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239"} err="failed to get container status \"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239\": rpc error: code = NotFound desc = could not find container \"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239\": container with ID starting with 21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.183382 4848 scope.go:117] "RemoveContainer" containerID="ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.183621 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\": container with ID starting with ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6 not found: ID does not exist" containerID="ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.183677 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6"} err="failed to get container status \"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\": rpc error: code = NotFound desc = could not find container \"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\": container with ID starting with ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.183697 4848 scope.go:117] "RemoveContainer" containerID="17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.183912 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\": container with ID starting with 17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907 not found: ID does not exist" containerID="17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.183965 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907"} err="failed to get container status \"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\": rpc error: code = NotFound desc = could not find container \"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\": container with ID starting with 17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.183980 4848 scope.go:117] "RemoveContainer" containerID="89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.184233 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\": container with ID starting with 89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591 not found: ID does not exist" containerID="89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.184276 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591"} err="failed to get container status \"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\": rpc error: code = NotFound desc = could not find container \"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\": container with ID starting with 89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.184304 4848 scope.go:117] "RemoveContainer" containerID="5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.184517 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\": container with ID starting with 5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db not found: ID does not exist" containerID="5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.184541 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db"} err="failed to get container status \"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\": rpc error: code = NotFound desc = could not find container \"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\": container with ID starting with 5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.184555 4848 scope.go:117] "RemoveContainer" containerID="4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.184738 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\": container with ID starting with 4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390 not found: ID does not exist" containerID="4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.184809 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390"} err="failed to get container status \"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\": rpc error: code = NotFound desc = could not find container \"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\": container with ID starting with 4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.184826 4848 scope.go:117] "RemoveContainer" containerID="a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.185097 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\": container with ID starting with a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9 not found: ID does not exist" containerID="a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.185146 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9"} err="failed to get container status \"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\": rpc error: code = NotFound desc = could not find container \"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\": container with ID starting with a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.185166 4848 scope.go:117] "RemoveContainer" containerID="9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.185990 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\": container with ID starting with 9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5 not found: ID does not exist" containerID="9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.186019 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5"} err="failed to get container status \"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\": rpc error: code = NotFound desc = could not find container \"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\": container with ID starting with 9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.186035 4848 scope.go:117] "RemoveContainer" containerID="f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594" Dec 04 14:00:54 crc kubenswrapper[4848]: E1204 14:00:54.190288 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\": container with ID starting with f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594 not found: ID does not exist" containerID="f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.190332 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594"} err="failed to get container status \"f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\": rpc error: code = NotFound desc = could not find container \"f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\": container with ID starting with f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.190357 4848 scope.go:117] "RemoveContainer" containerID="21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.190647 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239"} err="failed to get container status \"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239\": rpc error: code = NotFound desc = could not find container \"21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239\": container with ID starting with 21d0f5ec40798302da1953a8715566040ae36c53b759bdfc6e10a7a7dff21239 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.190682 4848 scope.go:117] "RemoveContainer" containerID="ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.190881 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6"} err="failed to get container status \"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\": rpc error: code = NotFound desc = could not find container \"ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6\": container with ID starting with ef0b42a40122447a66f1074a9f460eb506c4084a2bf87159770946e9d5b1ffd6 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.190901 4848 scope.go:117] "RemoveContainer" containerID="17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191120 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907"} err="failed to get container status \"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\": rpc error: code = NotFound desc = could not find container \"17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907\": container with ID starting with 17aa0bc23c2187c1eeb39c96d4cb44bbb69bb2d29a108c6c65aac4fab3949907 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191142 4848 scope.go:117] "RemoveContainer" containerID="89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191306 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591"} err="failed to get container status \"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\": rpc error: code = NotFound desc = could not find container \"89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591\": container with ID starting with 89f9822ec9b70c6b86489f07a6508342b996c07c881f38f4734d13bb2ebec591 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191326 4848 scope.go:117] "RemoveContainer" containerID="5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191481 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db"} err="failed to get container status \"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\": rpc error: code = NotFound desc = could not find container \"5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db\": container with ID starting with 5ab8d790206138b953e29eebb1763c2bc6b566b8578e17c613db257328ba13db not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191512 4848 scope.go:117] "RemoveContainer" containerID="4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191675 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390"} err="failed to get container status \"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\": rpc error: code = NotFound desc = could not find container \"4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390\": container with ID starting with 4d4da2241bd9b74238dee9c8bca17d85dc1ffc52bda62df0be715d85d2d64390 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191693 4848 scope.go:117] "RemoveContainer" containerID="a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191848 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9"} err="failed to get container status \"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\": rpc error: code = NotFound desc = could not find container \"a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9\": container with ID starting with a33a31eca75837d764b38a202f8c735f3e2385e7213bd783f94fc22daeca25b9 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.191866 4848 scope.go:117] "RemoveContainer" containerID="9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.192041 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5"} err="failed to get container status \"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\": rpc error: code = NotFound desc = could not find container \"9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5\": container with ID starting with 9ce89ed9d4993df79a9dd55750796a47ba2f07e566ab23d8b4db94405b1c77e5 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.192061 4848 scope.go:117] "RemoveContainer" containerID="f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.192215 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594"} err="failed to get container status \"f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\": rpc error: code = NotFound desc = could not find container \"f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594\": container with ID starting with f01182e183061394de2b4391a4dd03a34421b61b4fe7a1f6007cfe4d66707594 not found: ID does not exist" Dec 04 14:00:54 crc kubenswrapper[4848]: I1204 14:00:54.401366 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88de7a37-3579-4630-9aa7-64654e68a472" path="/var/lib/kubelet/pods/88de7a37-3579-4630-9aa7-64654e68a472/volumes" Dec 04 14:00:55 crc kubenswrapper[4848]: I1204 14:00:55.022359 4848 generic.go:334] "Generic (PLEG): container finished" podID="eadb015a-0f6b-451c-a9c4-f3ccd6a6927c" containerID="ac943c67f48bf2191bf585f3384370e578448c3baafa54ab580e0b239548502b" exitCode=0 Dec 04 14:00:55 crc kubenswrapper[4848]: I1204 14:00:55.022396 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerDied","Data":"ac943c67f48bf2191bf585f3384370e578448c3baafa54ab580e0b239548502b"} Dec 04 14:00:55 crc kubenswrapper[4848]: I1204 14:00:55.022418 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"cd4472ace58529c1946c4c704223a09ae4180eca91d4dd29c35463062e864b85"} Dec 04 14:00:56 crc kubenswrapper[4848]: I1204 14:00:56.041439 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"3f22f7658d98f2b005f999e6c1193e70850e371ec6ddec0eb6a9807f765c2205"} Dec 04 14:00:56 crc kubenswrapper[4848]: I1204 14:00:56.042081 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"c8334b3e312e6cda47dd87f052983470d88bf7447122596b8d46c3b491cdef21"} Dec 04 14:00:56 crc kubenswrapper[4848]: I1204 14:00:56.042097 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"111c6d47fd8aa7fbb7bda32e3347b9c1e61d7f4e2a1cd4d8a8dda20d234be63d"} Dec 04 14:00:57 crc kubenswrapper[4848]: I1204 14:00:57.050435 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"224a4b4996976e836adfc7da93f7908fb1d546067ba21add8484623bfa264c20"} Dec 04 14:00:57 crc kubenswrapper[4848]: I1204 14:00:57.050702 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"54659e88afdee2b2695f74fd40be5e09fe6b63dafd85e0c77edd85d7c05f6a54"} Dec 04 14:00:57 crc kubenswrapper[4848]: I1204 14:00:57.050713 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"bfc60eab4527f7dd0fde4e54e84fad6a8a8396d4f1e6c3ace831a070cfffaa4b"} Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.317421 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95"] Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.318470 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.320353 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.320681 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.322168 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-tz7qr" Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.466693 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwr5t\" (UniqueName: \"kubernetes.io/projected/41de33f3-4fb3-48fa-988b-cb6f059cab9e-kube-api-access-xwr5t\") pod \"obo-prometheus-operator-668cf9dfbb-mwx95\" (UID: \"41de33f3-4fb3-48fa-988b-cb6f059cab9e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.570848 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwr5t\" (UniqueName: \"kubernetes.io/projected/41de33f3-4fb3-48fa-988b-cb6f059cab9e-kube-api-access-xwr5t\") pod \"obo-prometheus-operator-668cf9dfbb-mwx95\" (UID: \"41de33f3-4fb3-48fa-988b-cb6f059cab9e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.595050 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwr5t\" (UniqueName: \"kubernetes.io/projected/41de33f3-4fb3-48fa-988b-cb6f059cab9e-kube-api-access-xwr5t\") pod \"obo-prometheus-operator-668cf9dfbb-mwx95\" (UID: \"41de33f3-4fb3-48fa-988b-cb6f059cab9e\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:00:59 crc kubenswrapper[4848]: I1204 14:00:59.632772 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:00:59 crc kubenswrapper[4848]: E1204 14:00:59.660468 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators_41de33f3-4fb3-48fa-988b-cb6f059cab9e_0(3f8c3ae3bd859340100049c77d42d620cebcd157a5ca3c9b22ec5710ad910a84): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:00:59 crc kubenswrapper[4848]: E1204 14:00:59.660735 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators_41de33f3-4fb3-48fa-988b-cb6f059cab9e_0(3f8c3ae3bd859340100049c77d42d620cebcd157a5ca3c9b22ec5710ad910a84): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:00:59 crc kubenswrapper[4848]: E1204 14:00:59.660756 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators_41de33f3-4fb3-48fa-988b-cb6f059cab9e_0(3f8c3ae3bd859340100049c77d42d620cebcd157a5ca3c9b22ec5710ad910a84): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:00:59 crc kubenswrapper[4848]: E1204 14:00:59.660794 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators(41de33f3-4fb3-48fa-988b-cb6f059cab9e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators(41de33f3-4fb3-48fa-988b-cb6f059cab9e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators_41de33f3-4fb3-48fa-988b-cb6f059cab9e_0(3f8c3ae3bd859340100049c77d42d620cebcd157a5ca3c9b22ec5710ad910a84): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" podUID="41de33f3-4fb3-48fa-988b-cb6f059cab9e" Dec 04 14:01:01 crc kubenswrapper[4848]: I1204 14:01:01.976790 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz"] Dec 04 14:01:01 crc kubenswrapper[4848]: I1204 14:01:01.978019 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:01 crc kubenswrapper[4848]: W1204 14:01:01.980474 4848 reflector.go:561] object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-brmcd": failed to list *v1.Secret: secrets "obo-prometheus-operator-admission-webhook-dockercfg-brmcd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-operators": no relationship found between node 'crc' and this object Dec 04 14:01:01 crc kubenswrapper[4848]: W1204 14:01:01.980518 4848 reflector.go:561] object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert": failed to list *v1.Secret: secrets "obo-prometheus-operator-admission-webhook-service-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-operators": no relationship found between node 'crc' and this object Dec 04 14:01:01 crc kubenswrapper[4848]: E1204 14:01:01.980529 4848 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators\"/\"obo-prometheus-operator-admission-webhook-dockercfg-brmcd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"obo-prometheus-operator-admission-webhook-dockercfg-brmcd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 14:01:01 crc kubenswrapper[4848]: E1204 14:01:01.980550 4848 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators\"/\"obo-prometheus-operator-admission-webhook-service-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"obo-prometheus-operator-admission-webhook-service-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 14:01:01 crc kubenswrapper[4848]: I1204 14:01:01.992648 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w"] Dec 04 14:01:01 crc kubenswrapper[4848]: I1204 14:01:01.993408 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.104110 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be2f3162-5720-4171-82c9-4df4a4244ec8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w\" (UID: \"be2f3162-5720-4171-82c9-4df4a4244ec8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.104243 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be2f3162-5720-4171-82c9-4df4a4244ec8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w\" (UID: \"be2f3162-5720-4171-82c9-4df4a4244ec8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.104281 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/202cc8e2-1429-410d-8c7f-36062221f810-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz\" (UID: \"202cc8e2-1429-410d-8c7f-36062221f810\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.104308 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/202cc8e2-1429-410d-8c7f-36062221f810-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz\" (UID: \"202cc8e2-1429-410d-8c7f-36062221f810\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.205316 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be2f3162-5720-4171-82c9-4df4a4244ec8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w\" (UID: \"be2f3162-5720-4171-82c9-4df4a4244ec8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.205386 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be2f3162-5720-4171-82c9-4df4a4244ec8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w\" (UID: \"be2f3162-5720-4171-82c9-4df4a4244ec8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.205411 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/202cc8e2-1429-410d-8c7f-36062221f810-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz\" (UID: \"202cc8e2-1429-410d-8c7f-36062221f810\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.205432 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/202cc8e2-1429-410d-8c7f-36062221f810-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz\" (UID: \"202cc8e2-1429-410d-8c7f-36062221f810\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:02 crc kubenswrapper[4848]: I1204 14:01:02.998488 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.010975 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/202cc8e2-1429-410d-8c7f-36062221f810-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz\" (UID: \"202cc8e2-1429-410d-8c7f-36062221f810\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.010939 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/be2f3162-5720-4171-82c9-4df4a4244ec8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w\" (UID: \"be2f3162-5720-4171-82c9-4df4a4244ec8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.011379 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/be2f3162-5720-4171-82c9-4df4a4244ec8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w\" (UID: \"be2f3162-5720-4171-82c9-4df4a4244ec8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.011761 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/202cc8e2-1429-410d-8c7f-36062221f810-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz\" (UID: \"202cc8e2-1429-410d-8c7f-36062221f810\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.239292 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-brmcd" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.239351 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.242088 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:03 crc kubenswrapper[4848]: E1204 14:01:03.284104 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators_202cc8e2-1429-410d-8c7f-36062221f810_0(ea5620750897aa9c21c12b56a0b30cdfdc9011697b3b575d760b930cc6fc99b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:03 crc kubenswrapper[4848]: E1204 14:01:03.284189 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators_202cc8e2-1429-410d-8c7f-36062221f810_0(ea5620750897aa9c21c12b56a0b30cdfdc9011697b3b575d760b930cc6fc99b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:03 crc kubenswrapper[4848]: E1204 14:01:03.284220 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators_202cc8e2-1429-410d-8c7f-36062221f810_0(ea5620750897aa9c21c12b56a0b30cdfdc9011697b3b575d760b930cc6fc99b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:03 crc kubenswrapper[4848]: E1204 14:01:03.284287 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators(202cc8e2-1429-410d-8c7f-36062221f810)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators(202cc8e2-1429-410d-8c7f-36062221f810)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators_202cc8e2-1429-410d-8c7f-36062221f810_0(ea5620750897aa9c21c12b56a0b30cdfdc9011697b3b575d760b930cc6fc99b4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" podUID="202cc8e2-1429-410d-8c7f-36062221f810" Dec 04 14:01:03 crc kubenswrapper[4848]: E1204 14:01:03.291809 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators_be2f3162-5720-4171-82c9-4df4a4244ec8_0(76085c5367ef903222aee47901411e6166e884da0e66aa9234bbd661f130f476): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:03 crc kubenswrapper[4848]: E1204 14:01:03.291922 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators_be2f3162-5720-4171-82c9-4df4a4244ec8_0(76085c5367ef903222aee47901411e6166e884da0e66aa9234bbd661f130f476): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:03 crc kubenswrapper[4848]: E1204 14:01:03.291980 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators_be2f3162-5720-4171-82c9-4df4a4244ec8_0(76085c5367ef903222aee47901411e6166e884da0e66aa9234bbd661f130f476): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:03 crc kubenswrapper[4848]: E1204 14:01:03.292055 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators(be2f3162-5720-4171-82c9-4df4a4244ec8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators(be2f3162-5720-4171-82c9-4df4a4244ec8)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators_be2f3162-5720-4171-82c9-4df4a4244ec8_0(76085c5367ef903222aee47901411e6166e884da0e66aa9234bbd661f130f476): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" podUID="be2f3162-5720-4171-82c9-4df4a4244ec8" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.659926 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-fl7j8"] Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.660708 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.662559 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-2pntt" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.663019 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.756732 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tn52h"] Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.757713 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.759686 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-22x5r" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.829467 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wthbw\" (UniqueName: \"kubernetes.io/projected/7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb-kube-api-access-wthbw\") pod \"observability-operator-d8bb48f5d-fl7j8\" (UID: \"7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb\") " pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.829768 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-fl7j8\" (UID: \"7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb\") " pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.931696 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wthbw\" (UniqueName: \"kubernetes.io/projected/7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb-kube-api-access-wthbw\") pod \"observability-operator-d8bb48f5d-fl7j8\" (UID: \"7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb\") " pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.931769 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-fl7j8\" (UID: \"7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb\") " pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.931810 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d6694eaf-b146-407c-95b2-c52ccc355a77-openshift-service-ca\") pod \"perses-operator-5446b9c989-tn52h\" (UID: \"d6694eaf-b146-407c-95b2-c52ccc355a77\") " pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.931848 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66lrx\" (UniqueName: \"kubernetes.io/projected/d6694eaf-b146-407c-95b2-c52ccc355a77-kube-api-access-66lrx\") pod \"perses-operator-5446b9c989-tn52h\" (UID: \"d6694eaf-b146-407c-95b2-c52ccc355a77\") " pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.938500 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-fl7j8\" (UID: \"7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb\") " pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.950545 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wthbw\" (UniqueName: \"kubernetes.io/projected/7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb-kube-api-access-wthbw\") pod \"observability-operator-d8bb48f5d-fl7j8\" (UID: \"7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb\") " pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:03 crc kubenswrapper[4848]: I1204 14:01:03.978997 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:04 crc kubenswrapper[4848]: E1204 14:01:04.001125 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fl7j8_openshift-operators_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb_0(b537c5ea12044e0ba4fdcb88d98001614b2711b0d5a492c60cb1ce52fb58ec2a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:04 crc kubenswrapper[4848]: E1204 14:01:04.002133 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fl7j8_openshift-operators_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb_0(b537c5ea12044e0ba4fdcb88d98001614b2711b0d5a492c60cb1ce52fb58ec2a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:04 crc kubenswrapper[4848]: E1204 14:01:04.002237 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fl7j8_openshift-operators_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb_0(b537c5ea12044e0ba4fdcb88d98001614b2711b0d5a492c60cb1ce52fb58ec2a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:04 crc kubenswrapper[4848]: E1204 14:01:04.002367 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-fl7j8_openshift-operators(7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-fl7j8_openshift-operators(7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fl7j8_openshift-operators_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb_0(b537c5ea12044e0ba4fdcb88d98001614b2711b0d5a492c60cb1ce52fb58ec2a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" podUID="7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb" Dec 04 14:01:04 crc kubenswrapper[4848]: I1204 14:01:04.033108 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d6694eaf-b146-407c-95b2-c52ccc355a77-openshift-service-ca\") pod \"perses-operator-5446b9c989-tn52h\" (UID: \"d6694eaf-b146-407c-95b2-c52ccc355a77\") " pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:04 crc kubenswrapper[4848]: I1204 14:01:04.033173 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66lrx\" (UniqueName: \"kubernetes.io/projected/d6694eaf-b146-407c-95b2-c52ccc355a77-kube-api-access-66lrx\") pod \"perses-operator-5446b9c989-tn52h\" (UID: \"d6694eaf-b146-407c-95b2-c52ccc355a77\") " pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:04 crc kubenswrapper[4848]: I1204 14:01:04.034029 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/d6694eaf-b146-407c-95b2-c52ccc355a77-openshift-service-ca\") pod \"perses-operator-5446b9c989-tn52h\" (UID: \"d6694eaf-b146-407c-95b2-c52ccc355a77\") " pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:04 crc kubenswrapper[4848]: I1204 14:01:04.048821 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66lrx\" (UniqueName: \"kubernetes.io/projected/d6694eaf-b146-407c-95b2-c52ccc355a77-kube-api-access-66lrx\") pod \"perses-operator-5446b9c989-tn52h\" (UID: \"d6694eaf-b146-407c-95b2-c52ccc355a77\") " pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:04 crc kubenswrapper[4848]: I1204 14:01:04.071836 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:04 crc kubenswrapper[4848]: E1204 14:01:04.093596 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tn52h_openshift-operators_d6694eaf-b146-407c-95b2-c52ccc355a77_0(6d69f87cd52c76dd983b1a0df84a6e28298d46b5492ee7a73e2f1bcdb44ab80d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:04 crc kubenswrapper[4848]: E1204 14:01:04.093654 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tn52h_openshift-operators_d6694eaf-b146-407c-95b2-c52ccc355a77_0(6d69f87cd52c76dd983b1a0df84a6e28298d46b5492ee7a73e2f1bcdb44ab80d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:04 crc kubenswrapper[4848]: E1204 14:01:04.093681 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tn52h_openshift-operators_d6694eaf-b146-407c-95b2-c52ccc355a77_0(6d69f87cd52c76dd983b1a0df84a6e28298d46b5492ee7a73e2f1bcdb44ab80d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:04 crc kubenswrapper[4848]: E1204 14:01:04.093730 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-tn52h_openshift-operators(d6694eaf-b146-407c-95b2-c52ccc355a77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-tn52h_openshift-operators(d6694eaf-b146-407c-95b2-c52ccc355a77)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tn52h_openshift-operators_d6694eaf-b146-407c-95b2-c52ccc355a77_0(6d69f87cd52c76dd983b1a0df84a6e28298d46b5492ee7a73e2f1bcdb44ab80d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-tn52h" podUID="d6694eaf-b146-407c-95b2-c52ccc355a77" Dec 04 14:01:04 crc kubenswrapper[4848]: I1204 14:01:04.097877 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"4b06cad272921efc4167f6dbce05c588f3873e6f23bcc1965ceb3e40491d0b10"} Dec 04 14:01:07 crc kubenswrapper[4848]: I1204 14:01:07.005785 4848 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.124474 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" event={"ID":"eadb015a-0f6b-451c-a9c4-f3ccd6a6927c","Type":"ContainerStarted","Data":"9f2ef1d5a441d57450f4fb1bb09f0959eacee4b16aa6c4e06571ef14614ae60f"} Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.125886 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.125900 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.125910 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.162675 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.195684 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.246924 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" podStartSLOduration=15.246907797 podStartE2EDuration="15.246907797s" podCreationTimestamp="2025-12-04 14:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:08.231837129 +0000 UTC m=+772.174333657" watchObservedRunningTime="2025-12-04 14:01:08.246907797 +0000 UTC m=+772.189404325" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.278141 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-fl7j8"] Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.278261 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.278681 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.301492 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tn52h"] Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.301597 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.301907 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.316960 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fl7j8_openshift-operators_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb_0(84470a517ee8427bd045d192423f6c1f0f72989b9a2c9d384ebdfebb7fc0c49a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.317029 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fl7j8_openshift-operators_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb_0(84470a517ee8427bd045d192423f6c1f0f72989b9a2c9d384ebdfebb7fc0c49a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.317050 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fl7j8_openshift-operators_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb_0(84470a517ee8427bd045d192423f6c1f0f72989b9a2c9d384ebdfebb7fc0c49a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.317098 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-fl7j8_openshift-operators(7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-fl7j8_openshift-operators(7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-fl7j8_openshift-operators_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb_0(84470a517ee8427bd045d192423f6c1f0f72989b9a2c9d384ebdfebb7fc0c49a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" podUID="7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.339997 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w"] Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.340109 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.340554 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.350959 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz"] Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.351095 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.351558 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.358646 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tn52h_openshift-operators_d6694eaf-b146-407c-95b2-c52ccc355a77_0(1fca6ea1679121e726321f767c4e87d5bb60103d230f52a8bc642fb3a7ee09ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.358712 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tn52h_openshift-operators_d6694eaf-b146-407c-95b2-c52ccc355a77_0(1fca6ea1679121e726321f767c4e87d5bb60103d230f52a8bc642fb3a7ee09ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.358737 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tn52h_openshift-operators_d6694eaf-b146-407c-95b2-c52ccc355a77_0(1fca6ea1679121e726321f767c4e87d5bb60103d230f52a8bc642fb3a7ee09ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.358781 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-tn52h_openshift-operators(d6694eaf-b146-407c-95b2-c52ccc355a77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-tn52h_openshift-operators(d6694eaf-b146-407c-95b2-c52ccc355a77)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tn52h_openshift-operators_d6694eaf-b146-407c-95b2-c52ccc355a77_0(1fca6ea1679121e726321f767c4e87d5bb60103d230f52a8bc642fb3a7ee09ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-tn52h" podUID="d6694eaf-b146-407c-95b2-c52ccc355a77" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.362195 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95"] Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.362317 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:01:08 crc kubenswrapper[4848]: I1204 14:01:08.362714 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.392139 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators_be2f3162-5720-4171-82c9-4df4a4244ec8_0(8c37e3e6d36d0ccc2ade97215b1eecd86cfe6f565c23efa272566431f17bd4df): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.392202 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators_be2f3162-5720-4171-82c9-4df4a4244ec8_0(8c37e3e6d36d0ccc2ade97215b1eecd86cfe6f565c23efa272566431f17bd4df): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.392225 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators_be2f3162-5720-4171-82c9-4df4a4244ec8_0(8c37e3e6d36d0ccc2ade97215b1eecd86cfe6f565c23efa272566431f17bd4df): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.392277 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators(be2f3162-5720-4171-82c9-4df4a4244ec8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators(be2f3162-5720-4171-82c9-4df4a4244ec8)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_openshift-operators_be2f3162-5720-4171-82c9-4df4a4244ec8_0(8c37e3e6d36d0ccc2ade97215b1eecd86cfe6f565c23efa272566431f17bd4df): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" podUID="be2f3162-5720-4171-82c9-4df4a4244ec8" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.424727 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators_202cc8e2-1429-410d-8c7f-36062221f810_0(2f110ac848e06ad5d894bd0d44b3162ca3cba020f3fe25c24a13b17f0e9ea35f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.424796 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators_202cc8e2-1429-410d-8c7f-36062221f810_0(2f110ac848e06ad5d894bd0d44b3162ca3cba020f3fe25c24a13b17f0e9ea35f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.424817 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators_202cc8e2-1429-410d-8c7f-36062221f810_0(2f110ac848e06ad5d894bd0d44b3162ca3cba020f3fe25c24a13b17f0e9ea35f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.424860 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators(202cc8e2-1429-410d-8c7f-36062221f810)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators(202cc8e2-1429-410d-8c7f-36062221f810)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_openshift-operators_202cc8e2-1429-410d-8c7f-36062221f810_0(2f110ac848e06ad5d894bd0d44b3162ca3cba020f3fe25c24a13b17f0e9ea35f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" podUID="202cc8e2-1429-410d-8c7f-36062221f810" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.436055 4848 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators_41de33f3-4fb3-48fa-988b-cb6f059cab9e_0(ded35d2759a76c9ef86f70335f50ca8847bd0cff07b935294399ff4cb64796e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.436120 4848 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators_41de33f3-4fb3-48fa-988b-cb6f059cab9e_0(ded35d2759a76c9ef86f70335f50ca8847bd0cff07b935294399ff4cb64796e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.436141 4848 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators_41de33f3-4fb3-48fa-988b-cb6f059cab9e_0(ded35d2759a76c9ef86f70335f50ca8847bd0cff07b935294399ff4cb64796e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:01:08 crc kubenswrapper[4848]: E1204 14:01:08.436186 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators(41de33f3-4fb3-48fa-988b-cb6f059cab9e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators(41de33f3-4fb3-48fa-988b-cb6f059cab9e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-mwx95_openshift-operators_41de33f3-4fb3-48fa-988b-cb6f059cab9e_0(ded35d2759a76c9ef86f70335f50ca8847bd0cff07b935294399ff4cb64796e8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" podUID="41de33f3-4fb3-48fa-988b-cb6f059cab9e" Dec 04 14:01:14 crc kubenswrapper[4848]: I1204 14:01:14.314302 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:01:14 crc kubenswrapper[4848]: I1204 14:01:14.314832 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:01:19 crc kubenswrapper[4848]: I1204 14:01:19.393116 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:19 crc kubenswrapper[4848]: I1204 14:01:19.394294 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" Dec 04 14:01:19 crc kubenswrapper[4848]: I1204 14:01:19.874799 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w"] Dec 04 14:01:19 crc kubenswrapper[4848]: W1204 14:01:19.882760 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe2f3162_5720_4171_82c9_4df4a4244ec8.slice/crio-59e7297a08a2b2794995ce0902377f073be77d80868a8fa1743401ce2b219de4 WatchSource:0}: Error finding container 59e7297a08a2b2794995ce0902377f073be77d80868a8fa1743401ce2b219de4: Status 404 returned error can't find the container with id 59e7297a08a2b2794995ce0902377f073be77d80868a8fa1743401ce2b219de4 Dec 04 14:01:20 crc kubenswrapper[4848]: I1204 14:01:20.196814 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" event={"ID":"be2f3162-5720-4171-82c9-4df4a4244ec8","Type":"ContainerStarted","Data":"59e7297a08a2b2794995ce0902377f073be77d80868a8fa1743401ce2b219de4"} Dec 04 14:01:20 crc kubenswrapper[4848]: I1204 14:01:20.393186 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:20 crc kubenswrapper[4848]: I1204 14:01:20.394071 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:20 crc kubenswrapper[4848]: I1204 14:01:20.843204 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tn52h"] Dec 04 14:01:21 crc kubenswrapper[4848]: I1204 14:01:21.205185 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-tn52h" event={"ID":"d6694eaf-b146-407c-95b2-c52ccc355a77","Type":"ContainerStarted","Data":"ff180b77398427e2fe6c08d74a26774f10d5795ed9ba7d3439123c769b696e2b"} Dec 04 14:01:21 crc kubenswrapper[4848]: I1204 14:01:21.393145 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:01:21 crc kubenswrapper[4848]: I1204 14:01:21.393654 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" Dec 04 14:01:21 crc kubenswrapper[4848]: I1204 14:01:21.622944 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95"] Dec 04 14:01:21 crc kubenswrapper[4848]: W1204 14:01:21.638216 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41de33f3_4fb3_48fa_988b_cb6f059cab9e.slice/crio-0f86707237cddcf3ba785e7b3847ea144e67a043035f3f653b4eb4c02595e55d WatchSource:0}: Error finding container 0f86707237cddcf3ba785e7b3847ea144e67a043035f3f653b4eb4c02595e55d: Status 404 returned error can't find the container with id 0f86707237cddcf3ba785e7b3847ea144e67a043035f3f653b4eb4c02595e55d Dec 04 14:01:22 crc kubenswrapper[4848]: I1204 14:01:22.212302 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" event={"ID":"41de33f3-4fb3-48fa-988b-cb6f059cab9e","Type":"ContainerStarted","Data":"0f86707237cddcf3ba785e7b3847ea144e67a043035f3f653b4eb4c02595e55d"} Dec 04 14:01:22 crc kubenswrapper[4848]: I1204 14:01:22.392668 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:22 crc kubenswrapper[4848]: I1204 14:01:22.393227 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" Dec 04 14:01:22 crc kubenswrapper[4848]: I1204 14:01:22.695023 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz"] Dec 04 14:01:23 crc kubenswrapper[4848]: I1204 14:01:23.224937 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" event={"ID":"202cc8e2-1429-410d-8c7f-36062221f810","Type":"ContainerStarted","Data":"ab86b1e2d0066c6c08e9b2536c80089316da8459b3ce9b8f645a7ef3997f1843"} Dec 04 14:01:23 crc kubenswrapper[4848]: I1204 14:01:23.393284 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:23 crc kubenswrapper[4848]: I1204 14:01:23.393766 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:23 crc kubenswrapper[4848]: I1204 14:01:23.882330 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-fl7j8"] Dec 04 14:01:23 crc kubenswrapper[4848]: W1204 14:01:23.895534 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7532e1ac_73b7_4cea_ab3f_eac0c9dc27eb.slice/crio-92310e31b22b67b743e2aa9e81ba3515efde92cfd1d60da9c56dd58c2059a4ad WatchSource:0}: Error finding container 92310e31b22b67b743e2aa9e81ba3515efde92cfd1d60da9c56dd58c2059a4ad: Status 404 returned error can't find the container with id 92310e31b22b67b743e2aa9e81ba3515efde92cfd1d60da9c56dd58c2059a4ad Dec 04 14:01:24 crc kubenswrapper[4848]: I1204 14:01:24.137219 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vjh9w" Dec 04 14:01:24 crc kubenswrapper[4848]: I1204 14:01:24.241480 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" event={"ID":"7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb","Type":"ContainerStarted","Data":"92310e31b22b67b743e2aa9e81ba3515efde92cfd1d60da9c56dd58c2059a4ad"} Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.162551 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w5m2w"] Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.166876 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.171465 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5m2w"] Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.337331 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw5jz\" (UniqueName: \"kubernetes.io/projected/aaf5b204-a875-4016-98e7-5a40c65a038e-kube-api-access-bw5jz\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.337474 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-utilities\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.337521 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-catalog-content\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.441216 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw5jz\" (UniqueName: \"kubernetes.io/projected/aaf5b204-a875-4016-98e7-5a40c65a038e-kube-api-access-bw5jz\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.441735 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-utilities\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.442318 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-catalog-content\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.442455 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-utilities\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.442466 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-catalog-content\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.462290 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw5jz\" (UniqueName: \"kubernetes.io/projected/aaf5b204-a875-4016-98e7-5a40c65a038e-kube-api-access-bw5jz\") pod \"redhat-marketplace-w5m2w\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:33 crc kubenswrapper[4848]: I1204 14:01:33.495474 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.140632 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5m2w"] Dec 04 14:01:35 crc kubenswrapper[4848]: W1204 14:01:35.155452 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaf5b204_a875_4016_98e7_5a40c65a038e.slice/crio-256680529cc560eb6d7b1f62ddbc676c1e429796e37411c4f532cb475d260256 WatchSource:0}: Error finding container 256680529cc560eb6d7b1f62ddbc676c1e429796e37411c4f532cb475d260256: Status 404 returned error can't find the container with id 256680529cc560eb6d7b1f62ddbc676c1e429796e37411c4f532cb475d260256 Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.306980 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" event={"ID":"be2f3162-5720-4171-82c9-4df4a4244ec8","Type":"ContainerStarted","Data":"c9255529547fa95233c950e24ed436eb5ebbbdc331b057ce4f8ff35fff8f4c6d"} Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.309375 4848 generic.go:334] "Generic (PLEG): container finished" podID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerID="bdeef6774ac85d9a27f7450bdfb8b03533dccacd45a9b2629f466c1a57b7ba06" exitCode=0 Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.309412 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5m2w" event={"ID":"aaf5b204-a875-4016-98e7-5a40c65a038e","Type":"ContainerDied","Data":"bdeef6774ac85d9a27f7450bdfb8b03533dccacd45a9b2629f466c1a57b7ba06"} Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.309426 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5m2w" event={"ID":"aaf5b204-a875-4016-98e7-5a40c65a038e","Type":"ContainerStarted","Data":"256680529cc560eb6d7b1f62ddbc676c1e429796e37411c4f532cb475d260256"} Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.313664 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-tn52h" event={"ID":"d6694eaf-b146-407c-95b2-c52ccc355a77","Type":"ContainerStarted","Data":"8bcbfb89be7fe56af40eff032c2ec88f0a7fdc8b8d9488a16984a8041f913cd8"} Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.313845 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.315636 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" event={"ID":"41de33f3-4fb3-48fa-988b-cb6f059cab9e","Type":"ContainerStarted","Data":"748d830f1cfd09381e4d0aa59071a2a51de74927d34ca6e31ecbcdc75141a52c"} Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.317991 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" event={"ID":"202cc8e2-1429-410d-8c7f-36062221f810","Type":"ContainerStarted","Data":"6e917df1af7c8993c6ced00e8f6354af1bfbeaf33a6a0a5ac8314b0374c8728a"} Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.319085 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" event={"ID":"7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb","Type":"ContainerStarted","Data":"9b5f82ecf7f65e1fc99d7bc9abaf6364ba66348a802c68dcc3b5db90158138ee"} Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.319288 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.320148 4848 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-fl7j8 container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.7:8081/healthz\": dial tcp 10.217.0.7:8081: connect: connection refused" start-of-body= Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.320199 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" podUID="7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.7:8081/healthz\": dial tcp 10.217.0.7:8081: connect: connection refused" Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.329943 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w" podStartSLOduration=19.322180788 podStartE2EDuration="34.329921936s" podCreationTimestamp="2025-12-04 14:01:01 +0000 UTC" firstStartedPulling="2025-12-04 14:01:19.88603997 +0000 UTC m=+783.828536498" lastFinishedPulling="2025-12-04 14:01:34.893781118 +0000 UTC m=+798.836277646" observedRunningTime="2025-12-04 14:01:35.325441887 +0000 UTC m=+799.267938415" watchObservedRunningTime="2025-12-04 14:01:35.329921936 +0000 UTC m=+799.272418464" Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.355534 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz" podStartSLOduration=22.17564088 podStartE2EDuration="34.355512571s" podCreationTimestamp="2025-12-04 14:01:01 +0000 UTC" firstStartedPulling="2025-12-04 14:01:22.716748896 +0000 UTC m=+786.659245424" lastFinishedPulling="2025-12-04 14:01:34.896620587 +0000 UTC m=+798.839117115" observedRunningTime="2025-12-04 14:01:35.350844736 +0000 UTC m=+799.293341264" watchObservedRunningTime="2025-12-04 14:01:35.355512571 +0000 UTC m=+799.298009099" Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.375912 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" podStartSLOduration=21.33403696 podStartE2EDuration="32.375896347s" podCreationTimestamp="2025-12-04 14:01:03 +0000 UTC" firstStartedPulling="2025-12-04 14:01:23.905572578 +0000 UTC m=+787.848069116" lastFinishedPulling="2025-12-04 14:01:34.947431965 +0000 UTC m=+798.889928503" observedRunningTime="2025-12-04 14:01:35.372666049 +0000 UTC m=+799.315162577" watchObservedRunningTime="2025-12-04 14:01:35.375896347 +0000 UTC m=+799.318392885" Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.419587 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-tn52h" podStartSLOduration=18.380979301 podStartE2EDuration="32.419567571s" podCreationTimestamp="2025-12-04 14:01:03 +0000 UTC" firstStartedPulling="2025-12-04 14:01:20.857689909 +0000 UTC m=+784.800186437" lastFinishedPulling="2025-12-04 14:01:34.896278179 +0000 UTC m=+798.838774707" observedRunningTime="2025-12-04 14:01:35.403764016 +0000 UTC m=+799.346260544" watchObservedRunningTime="2025-12-04 14:01:35.419567571 +0000 UTC m=+799.362064109" Dec 04 14:01:35 crc kubenswrapper[4848]: I1204 14:01:35.439403 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-mwx95" podStartSLOduration=23.198152857 podStartE2EDuration="36.439383374s" podCreationTimestamp="2025-12-04 14:00:59 +0000 UTC" firstStartedPulling="2025-12-04 14:01:21.645819477 +0000 UTC m=+785.588316005" lastFinishedPulling="2025-12-04 14:01:34.887049984 +0000 UTC m=+798.829546522" observedRunningTime="2025-12-04 14:01:35.431920752 +0000 UTC m=+799.374417280" watchObservedRunningTime="2025-12-04 14:01:35.439383374 +0000 UTC m=+799.381879902" Dec 04 14:01:36 crc kubenswrapper[4848]: I1204 14:01:36.329238 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5m2w" event={"ID":"aaf5b204-a875-4016-98e7-5a40c65a038e","Type":"ContainerStarted","Data":"503de82119aa74ff44bea988570c83644efdefbae08e3d309f44179c2e9e5ecf"} Dec 04 14:01:36 crc kubenswrapper[4848]: I1204 14:01:36.343271 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-fl7j8" Dec 04 14:01:37 crc kubenswrapper[4848]: I1204 14:01:37.336491 4848 generic.go:334] "Generic (PLEG): container finished" podID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerID="503de82119aa74ff44bea988570c83644efdefbae08e3d309f44179c2e9e5ecf" exitCode=0 Dec 04 14:01:37 crc kubenswrapper[4848]: I1204 14:01:37.336551 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5m2w" event={"ID":"aaf5b204-a875-4016-98e7-5a40c65a038e","Type":"ContainerDied","Data":"503de82119aa74ff44bea988570c83644efdefbae08e3d309f44179c2e9e5ecf"} Dec 04 14:01:38 crc kubenswrapper[4848]: I1204 14:01:38.344236 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5m2w" event={"ID":"aaf5b204-a875-4016-98e7-5a40c65a038e","Type":"ContainerStarted","Data":"903d24ccdd76ec8b653c78d0a94dd5d393a58785069ec05dd0a73df426d2d616"} Dec 04 14:01:38 crc kubenswrapper[4848]: I1204 14:01:38.362594 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w5m2w" podStartSLOduration=2.873518854 podStartE2EDuration="5.362574214s" podCreationTimestamp="2025-12-04 14:01:33 +0000 UTC" firstStartedPulling="2025-12-04 14:01:35.310521793 +0000 UTC m=+799.253018321" lastFinishedPulling="2025-12-04 14:01:37.799577153 +0000 UTC m=+801.742073681" observedRunningTime="2025-12-04 14:01:38.359256703 +0000 UTC m=+802.301753231" watchObservedRunningTime="2025-12-04 14:01:38.362574214 +0000 UTC m=+802.305070752" Dec 04 14:01:43 crc kubenswrapper[4848]: I1204 14:01:43.496639 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:43 crc kubenswrapper[4848]: I1204 14:01:43.498063 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:43 crc kubenswrapper[4848]: I1204 14:01:43.534114 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.074391 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-tn52h" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.314125 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.314214 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.314277 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.315157 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97feeafde380711ec8873168a8a64ca37c03edc0904564e6ca0465fe48907206"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.315256 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://97feeafde380711ec8873168a8a64ca37c03edc0904564e6ca0465fe48907206" gracePeriod=600 Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.434240 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.479152 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5m2w"] Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.800575 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r78j7"] Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.802057 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-r78j7" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.806432 4848 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-bq9f6" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.807016 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.807129 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.825221 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wcntc"] Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.826091 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-wcntc" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.829161 4848 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-28qdt" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.836452 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wcntc"] Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.842520 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-76vk7"] Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.843460 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.846037 4848 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-gbg2n" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.863234 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-76vk7"] Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.869258 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r78j7"] Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.909634 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzc27\" (UniqueName: \"kubernetes.io/projected/d708a779-4728-4230-ba68-02eb26b4dc72-kube-api-access-wzc27\") pod \"cert-manager-cainjector-7f985d654d-r78j7\" (UID: \"d708a779-4728-4230-ba68-02eb26b4dc72\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r78j7" Dec 04 14:01:44 crc kubenswrapper[4848]: I1204 14:01:44.909704 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc4bf\" (UniqueName: \"kubernetes.io/projected/327abad7-b085-4c01-8ecd-e287bda9f6ee-kube-api-access-jc4bf\") pod \"cert-manager-5b446d88c5-wcntc\" (UID: \"327abad7-b085-4c01-8ecd-e287bda9f6ee\") " pod="cert-manager/cert-manager-5b446d88c5-wcntc" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.010441 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzc27\" (UniqueName: \"kubernetes.io/projected/d708a779-4728-4230-ba68-02eb26b4dc72-kube-api-access-wzc27\") pod \"cert-manager-cainjector-7f985d654d-r78j7\" (UID: \"d708a779-4728-4230-ba68-02eb26b4dc72\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r78j7" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.010492 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc4bf\" (UniqueName: \"kubernetes.io/projected/327abad7-b085-4c01-8ecd-e287bda9f6ee-kube-api-access-jc4bf\") pod \"cert-manager-5b446d88c5-wcntc\" (UID: \"327abad7-b085-4c01-8ecd-e287bda9f6ee\") " pod="cert-manager/cert-manager-5b446d88c5-wcntc" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.010561 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9tkp\" (UniqueName: \"kubernetes.io/projected/7e550e0e-d6e6-4786-ae2b-46a4f1534877-kube-api-access-r9tkp\") pod \"cert-manager-webhook-5655c58dd6-76vk7\" (UID: \"7e550e0e-d6e6-4786-ae2b-46a4f1534877\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.028924 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzc27\" (UniqueName: \"kubernetes.io/projected/d708a779-4728-4230-ba68-02eb26b4dc72-kube-api-access-wzc27\") pod \"cert-manager-cainjector-7f985d654d-r78j7\" (UID: \"d708a779-4728-4230-ba68-02eb26b4dc72\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r78j7" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.031217 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc4bf\" (UniqueName: \"kubernetes.io/projected/327abad7-b085-4c01-8ecd-e287bda9f6ee-kube-api-access-jc4bf\") pod \"cert-manager-5b446d88c5-wcntc\" (UID: \"327abad7-b085-4c01-8ecd-e287bda9f6ee\") " pod="cert-manager/cert-manager-5b446d88c5-wcntc" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.111602 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9tkp\" (UniqueName: \"kubernetes.io/projected/7e550e0e-d6e6-4786-ae2b-46a4f1534877-kube-api-access-r9tkp\") pod \"cert-manager-webhook-5655c58dd6-76vk7\" (UID: \"7e550e0e-d6e6-4786-ae2b-46a4f1534877\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.120448 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-r78j7" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.139402 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9tkp\" (UniqueName: \"kubernetes.io/projected/7e550e0e-d6e6-4786-ae2b-46a4f1534877-kube-api-access-r9tkp\") pod \"cert-manager-webhook-5655c58dd6-76vk7\" (UID: \"7e550e0e-d6e6-4786-ae2b-46a4f1534877\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.141540 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-wcntc" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.155897 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.403473 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="97feeafde380711ec8873168a8a64ca37c03edc0904564e6ca0465fe48907206" exitCode=0 Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.403710 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"97feeafde380711ec8873168a8a64ca37c03edc0904564e6ca0465fe48907206"} Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.404061 4848 scope.go:117] "RemoveContainer" containerID="c3cebfe15e9ee005602e42c9fe3958ff03b1e9bc1b525474da30cb9fd280b5af" Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.413490 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r78j7"] Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.464608 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-wcntc"] Dec 04 14:01:45 crc kubenswrapper[4848]: I1204 14:01:45.742747 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-76vk7"] Dec 04 14:01:45 crc kubenswrapper[4848]: W1204 14:01:45.750626 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e550e0e_d6e6_4786_ae2b_46a4f1534877.slice/crio-257369ae8fb650b19ae0359306ab6b6e351a5d65920471d886b8322b34ffadf3 WatchSource:0}: Error finding container 257369ae8fb650b19ae0359306ab6b6e351a5d65920471d886b8322b34ffadf3: Status 404 returned error can't find the container with id 257369ae8fb650b19ae0359306ab6b6e351a5d65920471d886b8322b34ffadf3 Dec 04 14:01:46 crc kubenswrapper[4848]: I1204 14:01:46.432234 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-r78j7" event={"ID":"d708a779-4728-4230-ba68-02eb26b4dc72","Type":"ContainerStarted","Data":"0477ec551043f0d5e36069283b607ef0fbc5e7bf70be35744c996dd18b2d0cef"} Dec 04 14:01:46 crc kubenswrapper[4848]: I1204 14:01:46.434366 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-wcntc" event={"ID":"327abad7-b085-4c01-8ecd-e287bda9f6ee","Type":"ContainerStarted","Data":"fca30de5e3dc57d8db36a46eab9e7120d0ae2ce992cc9b3ce79ddc4d4f4fb066"} Dec 04 14:01:46 crc kubenswrapper[4848]: I1204 14:01:46.461444 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"f4368b11f240a04bf6f6e83cbceaec0f4a992be59424fb78d630b078c54c789d"} Dec 04 14:01:46 crc kubenswrapper[4848]: I1204 14:01:46.463462 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w5m2w" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerName="registry-server" containerID="cri-o://903d24ccdd76ec8b653c78d0a94dd5d393a58785069ec05dd0a73df426d2d616" gracePeriod=2 Dec 04 14:01:46 crc kubenswrapper[4848]: I1204 14:01:46.463747 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" event={"ID":"7e550e0e-d6e6-4786-ae2b-46a4f1534877","Type":"ContainerStarted","Data":"257369ae8fb650b19ae0359306ab6b6e351a5d65920471d886b8322b34ffadf3"} Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.482236 4848 generic.go:334] "Generic (PLEG): container finished" podID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerID="903d24ccdd76ec8b653c78d0a94dd5d393a58785069ec05dd0a73df426d2d616" exitCode=0 Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.483191 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5m2w" event={"ID":"aaf5b204-a875-4016-98e7-5a40c65a038e","Type":"ContainerDied","Data":"903d24ccdd76ec8b653c78d0a94dd5d393a58785069ec05dd0a73df426d2d616"} Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.560309 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.656543 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-catalog-content\") pod \"aaf5b204-a875-4016-98e7-5a40c65a038e\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.656596 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-utilities\") pod \"aaf5b204-a875-4016-98e7-5a40c65a038e\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.656705 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw5jz\" (UniqueName: \"kubernetes.io/projected/aaf5b204-a875-4016-98e7-5a40c65a038e-kube-api-access-bw5jz\") pod \"aaf5b204-a875-4016-98e7-5a40c65a038e\" (UID: \"aaf5b204-a875-4016-98e7-5a40c65a038e\") " Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.657749 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-utilities" (OuterVolumeSpecName: "utilities") pod "aaf5b204-a875-4016-98e7-5a40c65a038e" (UID: "aaf5b204-a875-4016-98e7-5a40c65a038e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.676391 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaf5b204-a875-4016-98e7-5a40c65a038e-kube-api-access-bw5jz" (OuterVolumeSpecName: "kube-api-access-bw5jz") pod "aaf5b204-a875-4016-98e7-5a40c65a038e" (UID: "aaf5b204-a875-4016-98e7-5a40c65a038e"). InnerVolumeSpecName "kube-api-access-bw5jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.692352 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aaf5b204-a875-4016-98e7-5a40c65a038e" (UID: "aaf5b204-a875-4016-98e7-5a40c65a038e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.758497 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw5jz\" (UniqueName: \"kubernetes.io/projected/aaf5b204-a875-4016-98e7-5a40c65a038e-kube-api-access-bw5jz\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.758531 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:47 crc kubenswrapper[4848]: I1204 14:01:47.758542 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaf5b204-a875-4016-98e7-5a40c65a038e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:48 crc kubenswrapper[4848]: I1204 14:01:48.492799 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5m2w" event={"ID":"aaf5b204-a875-4016-98e7-5a40c65a038e","Type":"ContainerDied","Data":"256680529cc560eb6d7b1f62ddbc676c1e429796e37411c4f532cb475d260256"} Dec 04 14:01:48 crc kubenswrapper[4848]: I1204 14:01:48.492863 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5m2w" Dec 04 14:01:48 crc kubenswrapper[4848]: I1204 14:01:48.493169 4848 scope.go:117] "RemoveContainer" containerID="903d24ccdd76ec8b653c78d0a94dd5d393a58785069ec05dd0a73df426d2d616" Dec 04 14:01:48 crc kubenswrapper[4848]: I1204 14:01:48.521069 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5m2w"] Dec 04 14:01:48 crc kubenswrapper[4848]: I1204 14:01:48.526273 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5m2w"] Dec 04 14:01:50 crc kubenswrapper[4848]: I1204 14:01:50.401780 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" path="/var/lib/kubelet/pods/aaf5b204-a875-4016-98e7-5a40c65a038e/volumes" Dec 04 14:01:55 crc kubenswrapper[4848]: I1204 14:01:55.242014 4848 scope.go:117] "RemoveContainer" containerID="503de82119aa74ff44bea988570c83644efdefbae08e3d309f44179c2e9e5ecf" Dec 04 14:01:55 crc kubenswrapper[4848]: I1204 14:01:55.737116 4848 scope.go:117] "RemoveContainer" containerID="bdeef6774ac85d9a27f7450bdfb8b03533dccacd45a9b2629f466c1a57b7ba06" Dec 04 14:01:56 crc kubenswrapper[4848]: I1204 14:01:56.554165 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-r78j7" event={"ID":"d708a779-4728-4230-ba68-02eb26b4dc72","Type":"ContainerStarted","Data":"fdb4a8c6428114f09ebf5b9a084a7f72b3dcafcaf311a089b5e69868e25e9c93"} Dec 04 14:01:56 crc kubenswrapper[4848]: I1204 14:01:56.555654 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-wcntc" event={"ID":"327abad7-b085-4c01-8ecd-e287bda9f6ee","Type":"ContainerStarted","Data":"3be783abbc63affcaa99e613b33aa65c0d9c4a3972c1b582f3f812d3cce52f0e"} Dec 04 14:01:56 crc kubenswrapper[4848]: I1204 14:01:56.556871 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" event={"ID":"7e550e0e-d6e6-4786-ae2b-46a4f1534877","Type":"ContainerStarted","Data":"d046b4da65f29631147abe4a17e4c855734a34a0c00f9287822f3d3e1d595909"} Dec 04 14:01:56 crc kubenswrapper[4848]: I1204 14:01:56.556926 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" Dec 04 14:01:56 crc kubenswrapper[4848]: I1204 14:01:56.568593 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-r78j7" podStartSLOduration=2.09916859 podStartE2EDuration="12.568573277s" podCreationTimestamp="2025-12-04 14:01:44 +0000 UTC" firstStartedPulling="2025-12-04 14:01:45.417157489 +0000 UTC m=+809.359654017" lastFinishedPulling="2025-12-04 14:01:55.886562176 +0000 UTC m=+819.829058704" observedRunningTime="2025-12-04 14:01:56.567338227 +0000 UTC m=+820.509834745" watchObservedRunningTime="2025-12-04 14:01:56.568573277 +0000 UTC m=+820.511069815" Dec 04 14:01:56 crc kubenswrapper[4848]: I1204 14:01:56.586688 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-wcntc" podStartSLOduration=2.115515067 podStartE2EDuration="12.586668388s" podCreationTimestamp="2025-12-04 14:01:44 +0000 UTC" firstStartedPulling="2025-12-04 14:01:45.489542882 +0000 UTC m=+809.432039410" lastFinishedPulling="2025-12-04 14:01:55.960696203 +0000 UTC m=+819.903192731" observedRunningTime="2025-12-04 14:01:56.583247155 +0000 UTC m=+820.525743693" watchObservedRunningTime="2025-12-04 14:01:56.586668388 +0000 UTC m=+820.529164926" Dec 04 14:01:56 crc kubenswrapper[4848]: I1204 14:01:56.603032 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" podStartSLOduration=2.390286455 podStartE2EDuration="12.603002887s" podCreationTimestamp="2025-12-04 14:01:44 +0000 UTC" firstStartedPulling="2025-12-04 14:01:45.752407719 +0000 UTC m=+809.694904247" lastFinishedPulling="2025-12-04 14:01:55.965124151 +0000 UTC m=+819.907620679" observedRunningTime="2025-12-04 14:01:56.599915982 +0000 UTC m=+820.542412510" watchObservedRunningTime="2025-12-04 14:01:56.603002887 +0000 UTC m=+820.545499425" Dec 04 14:02:05 crc kubenswrapper[4848]: I1204 14:02:05.158860 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-76vk7" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.870310 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk"] Dec 04 14:02:28 crc kubenswrapper[4848]: E1204 14:02:28.871122 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerName="registry-server" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.871138 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerName="registry-server" Dec 04 14:02:28 crc kubenswrapper[4848]: E1204 14:02:28.871164 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerName="extract-content" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.871171 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerName="extract-content" Dec 04 14:02:28 crc kubenswrapper[4848]: E1204 14:02:28.871184 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerName="extract-utilities" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.871215 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerName="extract-utilities" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.871363 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf5b204-a875-4016-98e7-5a40c65a038e" containerName="registry-server" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.872434 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.874704 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.885438 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk"] Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.966360 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.966785 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmfmr\" (UniqueName: \"kubernetes.io/projected/670575a5-b7cb-43bb-bfa9-2cf7666f3449-kube-api-access-gmfmr\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:28 crc kubenswrapper[4848]: I1204 14:02:28.966820 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.040258 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k"] Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.041502 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.053558 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k"] Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.068678 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.068828 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.068871 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmfmr\" (UniqueName: \"kubernetes.io/projected/670575a5-b7cb-43bb-bfa9-2cf7666f3449-kube-api-access-gmfmr\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.068912 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.069029 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vf9d\" (UniqueName: \"kubernetes.io/projected/2748bd61-6b69-4c1f-95dd-f8465b1b542b-kube-api-access-7vf9d\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.069087 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.069917 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.069919 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.091287 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmfmr\" (UniqueName: \"kubernetes.io/projected/670575a5-b7cb-43bb-bfa9-2cf7666f3449-kube-api-access-gmfmr\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.170108 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vf9d\" (UniqueName: \"kubernetes.io/projected/2748bd61-6b69-4c1f-95dd-f8465b1b542b-kube-api-access-7vf9d\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.170234 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.170288 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.170785 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.171005 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.187023 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vf9d\" (UniqueName: \"kubernetes.io/projected/2748bd61-6b69-4c1f-95dd-f8465b1b542b-kube-api-access-7vf9d\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.236119 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.356620 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.450429 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk"] Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.802906 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k"] Dec 04 14:02:29 crc kubenswrapper[4848]: I1204 14:02:29.807900 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" event={"ID":"670575a5-b7cb-43bb-bfa9-2cf7666f3449","Type":"ContainerStarted","Data":"aa600a53e4135b4b531b6930cb4e8ed68b448c43176b3855b2c61c59e24b4a17"} Dec 04 14:02:29 crc kubenswrapper[4848]: W1204 14:02:29.817030 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2748bd61_6b69_4c1f_95dd_f8465b1b542b.slice/crio-727be3b4df06e09184df45b4a2914e0ea1c2c3b3ac29cde65f76971c72b2fe00 WatchSource:0}: Error finding container 727be3b4df06e09184df45b4a2914e0ea1c2c3b3ac29cde65f76971c72b2fe00: Status 404 returned error can't find the container with id 727be3b4df06e09184df45b4a2914e0ea1c2c3b3ac29cde65f76971c72b2fe00 Dec 04 14:02:30 crc kubenswrapper[4848]: I1204 14:02:30.816194 4848 generic.go:334] "Generic (PLEG): container finished" podID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerID="20495cbccc18d353f2df9e572cb72ddb5f49cfe301f93defcdb223fd4be25e33" exitCode=0 Dec 04 14:02:30 crc kubenswrapper[4848]: I1204 14:02:30.816314 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" event={"ID":"2748bd61-6b69-4c1f-95dd-f8465b1b542b","Type":"ContainerDied","Data":"20495cbccc18d353f2df9e572cb72ddb5f49cfe301f93defcdb223fd4be25e33"} Dec 04 14:02:30 crc kubenswrapper[4848]: I1204 14:02:30.816348 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" event={"ID":"2748bd61-6b69-4c1f-95dd-f8465b1b542b","Type":"ContainerStarted","Data":"727be3b4df06e09184df45b4a2914e0ea1c2c3b3ac29cde65f76971c72b2fe00"} Dec 04 14:02:30 crc kubenswrapper[4848]: I1204 14:02:30.818898 4848 generic.go:334] "Generic (PLEG): container finished" podID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerID="d4fb02d8cc013bd6de51b656c79a5b3eea2d9bd5e43fbf21c763b9631514a2a5" exitCode=0 Dec 04 14:02:30 crc kubenswrapper[4848]: I1204 14:02:30.818984 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" event={"ID":"670575a5-b7cb-43bb-bfa9-2cf7666f3449","Type":"ContainerDied","Data":"d4fb02d8cc013bd6de51b656c79a5b3eea2d9bd5e43fbf21c763b9631514a2a5"} Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.590906 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fg52v"] Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.592822 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.603658 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fg52v"] Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.615498 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-catalog-content\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.615613 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-utilities\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.615711 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7wdp\" (UniqueName: \"kubernetes.io/projected/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-kube-api-access-l7wdp\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.716464 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-utilities\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.716781 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7wdp\" (UniqueName: \"kubernetes.io/projected/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-kube-api-access-l7wdp\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.716861 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-catalog-content\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.717323 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-catalog-content\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.717393 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-utilities\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.748970 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7wdp\" (UniqueName: \"kubernetes.io/projected/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-kube-api-access-l7wdp\") pod \"redhat-operators-fg52v\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.832672 4848 generic.go:334] "Generic (PLEG): container finished" podID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerID="11229d718f922730f4d6f71c48c95392746e8d494f94291d117b5286f5d3b219" exitCode=0 Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.832738 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" event={"ID":"2748bd61-6b69-4c1f-95dd-f8465b1b542b","Type":"ContainerDied","Data":"11229d718f922730f4d6f71c48c95392746e8d494f94291d117b5286f5d3b219"} Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.835145 4848 generic.go:334] "Generic (PLEG): container finished" podID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerID="a66221c587fc73e70d6ba51430e5bb533cb5d27087782c525c50168a458a875f" exitCode=0 Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.835182 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" event={"ID":"670575a5-b7cb-43bb-bfa9-2cf7666f3449","Type":"ContainerDied","Data":"a66221c587fc73e70d6ba51430e5bb533cb5d27087782c525c50168a458a875f"} Dec 04 14:02:32 crc kubenswrapper[4848]: I1204 14:02:32.920573 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:33 crc kubenswrapper[4848]: I1204 14:02:33.132776 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fg52v"] Dec 04 14:02:33 crc kubenswrapper[4848]: I1204 14:02:33.842399 4848 generic.go:334] "Generic (PLEG): container finished" podID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerID="2ee74b809f49253ca2a8994ab42fa1ec1ff7cbc21357ea4017269c1b2777e224" exitCode=0 Dec 04 14:02:33 crc kubenswrapper[4848]: I1204 14:02:33.842453 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" event={"ID":"670575a5-b7cb-43bb-bfa9-2cf7666f3449","Type":"ContainerDied","Data":"2ee74b809f49253ca2a8994ab42fa1ec1ff7cbc21357ea4017269c1b2777e224"} Dec 04 14:02:33 crc kubenswrapper[4848]: I1204 14:02:33.844080 4848 generic.go:334] "Generic (PLEG): container finished" podID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerID="a142fa8f6fb5eec08276065d8446f0aa69622cae54846ae626c2b8816383677f" exitCode=0 Dec 04 14:02:33 crc kubenswrapper[4848]: I1204 14:02:33.844118 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fg52v" event={"ID":"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf","Type":"ContainerDied","Data":"a142fa8f6fb5eec08276065d8446f0aa69622cae54846ae626c2b8816383677f"} Dec 04 14:02:33 crc kubenswrapper[4848]: I1204 14:02:33.844135 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fg52v" event={"ID":"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf","Type":"ContainerStarted","Data":"2d8f82146887cccd5c6aa92c7e3562569d39ef9c946f4fa5ca58d4268d2c46a3"} Dec 04 14:02:33 crc kubenswrapper[4848]: I1204 14:02:33.848100 4848 generic.go:334] "Generic (PLEG): container finished" podID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerID="13fcdbed43739495ab3d3558b59533c0615df769db21e4f9c142c0430dade388" exitCode=0 Dec 04 14:02:33 crc kubenswrapper[4848]: I1204 14:02:33.848134 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" event={"ID":"2748bd61-6b69-4c1f-95dd-f8465b1b542b","Type":"ContainerDied","Data":"13fcdbed43739495ab3d3558b59533c0615df769db21e4f9c142c0430dade388"} Dec 04 14:02:34 crc kubenswrapper[4848]: I1204 14:02:34.859109 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fg52v" event={"ID":"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf","Type":"ContainerStarted","Data":"e79615966c2c5527fbe43dfd762cd4bbb7df2e70cce109957ae826a93c316168"} Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.168720 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.170261 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.354800 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-bundle\") pod \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.354995 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vf9d\" (UniqueName: \"kubernetes.io/projected/2748bd61-6b69-4c1f-95dd-f8465b1b542b-kube-api-access-7vf9d\") pod \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.355049 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmfmr\" (UniqueName: \"kubernetes.io/projected/670575a5-b7cb-43bb-bfa9-2cf7666f3449-kube-api-access-gmfmr\") pod \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.355074 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-util\") pod \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\" (UID: \"670575a5-b7cb-43bb-bfa9-2cf7666f3449\") " Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.355118 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-bundle\") pod \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.355148 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-util\") pod \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\" (UID: \"2748bd61-6b69-4c1f-95dd-f8465b1b542b\") " Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.355781 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-bundle" (OuterVolumeSpecName: "bundle") pod "670575a5-b7cb-43bb-bfa9-2cf7666f3449" (UID: "670575a5-b7cb-43bb-bfa9-2cf7666f3449"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.356360 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-bundle" (OuterVolumeSpecName: "bundle") pod "2748bd61-6b69-4c1f-95dd-f8465b1b542b" (UID: "2748bd61-6b69-4c1f-95dd-f8465b1b542b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.360575 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2748bd61-6b69-4c1f-95dd-f8465b1b542b-kube-api-access-7vf9d" (OuterVolumeSpecName: "kube-api-access-7vf9d") pod "2748bd61-6b69-4c1f-95dd-f8465b1b542b" (UID: "2748bd61-6b69-4c1f-95dd-f8465b1b542b"). InnerVolumeSpecName "kube-api-access-7vf9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.360708 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/670575a5-b7cb-43bb-bfa9-2cf7666f3449-kube-api-access-gmfmr" (OuterVolumeSpecName: "kube-api-access-gmfmr") pod "670575a5-b7cb-43bb-bfa9-2cf7666f3449" (UID: "670575a5-b7cb-43bb-bfa9-2cf7666f3449"). InnerVolumeSpecName "kube-api-access-gmfmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.385436 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-util" (OuterVolumeSpecName: "util") pod "670575a5-b7cb-43bb-bfa9-2cf7666f3449" (UID: "670575a5-b7cb-43bb-bfa9-2cf7666f3449"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.395118 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-util" (OuterVolumeSpecName: "util") pod "2748bd61-6b69-4c1f-95dd-f8465b1b542b" (UID: "2748bd61-6b69-4c1f-95dd-f8465b1b542b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.457357 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vf9d\" (UniqueName: \"kubernetes.io/projected/2748bd61-6b69-4c1f-95dd-f8465b1b542b-kube-api-access-7vf9d\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.457391 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmfmr\" (UniqueName: \"kubernetes.io/projected/670575a5-b7cb-43bb-bfa9-2cf7666f3449-kube-api-access-gmfmr\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.457401 4848 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.457411 4848 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.457419 4848 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2748bd61-6b69-4c1f-95dd-f8465b1b542b-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.457426 4848 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/670575a5-b7cb-43bb-bfa9-2cf7666f3449-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.866435 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" event={"ID":"2748bd61-6b69-4c1f-95dd-f8465b1b542b","Type":"ContainerDied","Data":"727be3b4df06e09184df45b4a2914e0ea1c2c3b3ac29cde65f76971c72b2fe00"} Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.866623 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="727be3b4df06e09184df45b4a2914e0ea1c2c3b3ac29cde65f76971c72b2fe00" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.866477 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.868978 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" event={"ID":"670575a5-b7cb-43bb-bfa9-2cf7666f3449","Type":"ContainerDied","Data":"aa600a53e4135b4b531b6930cb4e8ed68b448c43176b3855b2c61c59e24b4a17"} Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.869020 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa600a53e4135b4b531b6930cb4e8ed68b448c43176b3855b2c61c59e24b4a17" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.869111 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk" Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.870717 4848 generic.go:334] "Generic (PLEG): container finished" podID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerID="e79615966c2c5527fbe43dfd762cd4bbb7df2e70cce109957ae826a93c316168" exitCode=0 Dec 04 14:02:35 crc kubenswrapper[4848]: I1204 14:02:35.870781 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fg52v" event={"ID":"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf","Type":"ContainerDied","Data":"e79615966c2c5527fbe43dfd762cd4bbb7df2e70cce109957ae826a93c316168"} Dec 04 14:02:36 crc kubenswrapper[4848]: I1204 14:02:36.878470 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fg52v" event={"ID":"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf","Type":"ContainerStarted","Data":"aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49"} Dec 04 14:02:36 crc kubenswrapper[4848]: I1204 14:02:36.901318 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fg52v" podStartSLOduration=2.38326454 podStartE2EDuration="4.901294617s" podCreationTimestamp="2025-12-04 14:02:32 +0000 UTC" firstStartedPulling="2025-12-04 14:02:33.84565472 +0000 UTC m=+857.788151248" lastFinishedPulling="2025-12-04 14:02:36.363684797 +0000 UTC m=+860.306181325" observedRunningTime="2025-12-04 14:02:36.89568832 +0000 UTC m=+860.838184868" watchObservedRunningTime="2025-12-04 14:02:36.901294617 +0000 UTC m=+860.843791145" Dec 04 14:02:42 crc kubenswrapper[4848]: I1204 14:02:42.922031 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:42 crc kubenswrapper[4848]: I1204 14:02:42.922483 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:42 crc kubenswrapper[4848]: I1204 14:02:42.986472 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:43 crc kubenswrapper[4848]: I1204 14:02:43.967235 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:45 crc kubenswrapper[4848]: I1204 14:02:45.578169 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fg52v"] Dec 04 14:02:45 crc kubenswrapper[4848]: I1204 14:02:45.939579 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fg52v" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="registry-server" containerID="cri-o://aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49" gracePeriod=2 Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.089221 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg"] Dec 04 14:02:47 crc kubenswrapper[4848]: E1204 14:02:47.089828 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerName="pull" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.089843 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerName="pull" Dec 04 14:02:47 crc kubenswrapper[4848]: E1204 14:02:47.089860 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerName="pull" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.089869 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerName="pull" Dec 04 14:02:47 crc kubenswrapper[4848]: E1204 14:02:47.089887 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerName="extract" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.089895 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerName="extract" Dec 04 14:02:47 crc kubenswrapper[4848]: E1204 14:02:47.089908 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerName="util" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.089915 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerName="util" Dec 04 14:02:47 crc kubenswrapper[4848]: E1204 14:02:47.089927 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerName="util" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.089934 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerName="util" Dec 04 14:02:47 crc kubenswrapper[4848]: E1204 14:02:47.089971 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerName="extract" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.089979 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerName="extract" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.090133 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="2748bd61-6b69-4c1f-95dd-f8465b1b542b" containerName="extract" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.090150 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="670575a5-b7cb-43bb-bfa9-2cf7666f3449" containerName="extract" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.090984 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.092803 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-hd8v4" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.093353 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.093583 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.096731 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.097838 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.098262 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.108532 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg"] Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.233921 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6dmg\" (UniqueName: \"kubernetes.io/projected/461d1ece-26ff-40ed-a92c-f21300adeec7-kube-api-access-h6dmg\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.234599 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/461d1ece-26ff-40ed-a92c-f21300adeec7-manager-config\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.234634 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-webhook-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.234655 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.234688 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-apiservice-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.335667 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6dmg\" (UniqueName: \"kubernetes.io/projected/461d1ece-26ff-40ed-a92c-f21300adeec7-kube-api-access-h6dmg\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.335727 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/461d1ece-26ff-40ed-a92c-f21300adeec7-manager-config\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.335765 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-webhook-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.335795 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.335823 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-apiservice-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.336731 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/461d1ece-26ff-40ed-a92c-f21300adeec7-manager-config\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.341405 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.341436 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-apiservice-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.341911 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/461d1ece-26ff-40ed-a92c-f21300adeec7-webhook-cert\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.366278 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6dmg\" (UniqueName: \"kubernetes.io/projected/461d1ece-26ff-40ed-a92c-f21300adeec7-kube-api-access-h6dmg\") pod \"loki-operator-controller-manager-7ddb79586-x6vmg\" (UID: \"461d1ece-26ff-40ed-a92c-f21300adeec7\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.404355 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.621928 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg"] Dec 04 14:02:47 crc kubenswrapper[4848]: W1204 14:02:47.642441 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod461d1ece_26ff_40ed_a92c_f21300adeec7.slice/crio-94a94f4a0a1eecca5b2d7abd847fd808e73fd160adc25439fc876e9012d564f6 WatchSource:0}: Error finding container 94a94f4a0a1eecca5b2d7abd847fd808e73fd160adc25439fc876e9012d564f6: Status 404 returned error can't find the container with id 94a94f4a0a1eecca5b2d7abd847fd808e73fd160adc25439fc876e9012d564f6 Dec 04 14:02:47 crc kubenswrapper[4848]: I1204 14:02:47.959084 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" event={"ID":"461d1ece-26ff-40ed-a92c-f21300adeec7","Type":"ContainerStarted","Data":"94a94f4a0a1eecca5b2d7abd847fd808e73fd160adc25439fc876e9012d564f6"} Dec 04 14:02:48 crc kubenswrapper[4848]: I1204 14:02:48.837890 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lgmcj"] Dec 04 14:02:48 crc kubenswrapper[4848]: I1204 14:02:48.839349 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-lgmcj" Dec 04 14:02:48 crc kubenswrapper[4848]: I1204 14:02:48.844489 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-dqh8r" Dec 04 14:02:48 crc kubenswrapper[4848]: I1204 14:02:48.853033 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 04 14:02:48 crc kubenswrapper[4848]: I1204 14:02:48.853832 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lgmcj"] Dec 04 14:02:48 crc kubenswrapper[4848]: I1204 14:02:48.854278 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 04 14:02:48 crc kubenswrapper[4848]: I1204 14:02:48.963976 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5trz\" (UniqueName: \"kubernetes.io/projected/5d3f5a36-9e68-4f3d-94b8-8cb59192211b-kube-api-access-k5trz\") pod \"cluster-logging-operator-ff9846bd-lgmcj\" (UID: \"5d3f5a36-9e68-4f3d-94b8-8cb59192211b\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lgmcj" Dec 04 14:02:49 crc kubenswrapper[4848]: I1204 14:02:49.065265 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5trz\" (UniqueName: \"kubernetes.io/projected/5d3f5a36-9e68-4f3d-94b8-8cb59192211b-kube-api-access-k5trz\") pod \"cluster-logging-operator-ff9846bd-lgmcj\" (UID: \"5d3f5a36-9e68-4f3d-94b8-8cb59192211b\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lgmcj" Dec 04 14:02:49 crc kubenswrapper[4848]: I1204 14:02:49.083645 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5trz\" (UniqueName: \"kubernetes.io/projected/5d3f5a36-9e68-4f3d-94b8-8cb59192211b-kube-api-access-k5trz\") pod \"cluster-logging-operator-ff9846bd-lgmcj\" (UID: \"5d3f5a36-9e68-4f3d-94b8-8cb59192211b\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-lgmcj" Dec 04 14:02:49 crc kubenswrapper[4848]: I1204 14:02:49.160862 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-lgmcj" Dec 04 14:02:49 crc kubenswrapper[4848]: I1204 14:02:49.593312 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-lgmcj"] Dec 04 14:02:49 crc kubenswrapper[4848]: W1204 14:02:49.601891 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d3f5a36_9e68_4f3d_94b8_8cb59192211b.slice/crio-d83b8a36d660b7d878815b22ea0e07f81e0d17c32858ec0ab10271a84b87ad40 WatchSource:0}: Error finding container d83b8a36d660b7d878815b22ea0e07f81e0d17c32858ec0ab10271a84b87ad40: Status 404 returned error can't find the container with id d83b8a36d660b7d878815b22ea0e07f81e0d17c32858ec0ab10271a84b87ad40 Dec 04 14:02:52 crc kubenswrapper[4848]: I1204 14:02:52.667983 4848 generic.go:334] "Generic (PLEG): container finished" podID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerID="aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49" exitCode=0 Dec 04 14:02:52 crc kubenswrapper[4848]: I1204 14:02:52.668089 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fg52v" event={"ID":"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf","Type":"ContainerDied","Data":"aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49"} Dec 04 14:02:52 crc kubenswrapper[4848]: I1204 14:02:52.669592 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-lgmcj" event={"ID":"5d3f5a36-9e68-4f3d-94b8-8cb59192211b","Type":"ContainerStarted","Data":"d83b8a36d660b7d878815b22ea0e07f81e0d17c32858ec0ab10271a84b87ad40"} Dec 04 14:02:52 crc kubenswrapper[4848]: E1204 14:02:52.922337 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49 is running failed: container process not found" containerID="aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:02:52 crc kubenswrapper[4848]: E1204 14:02:52.922845 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49 is running failed: container process not found" containerID="aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:02:52 crc kubenswrapper[4848]: E1204 14:02:52.923271 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49 is running failed: container process not found" containerID="aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:02:52 crc kubenswrapper[4848]: E1204 14:02:52.923339 4848 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-fg52v" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="registry-server" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.200433 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.375449 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-utilities\") pod \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.375618 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7wdp\" (UniqueName: \"kubernetes.io/projected/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-kube-api-access-l7wdp\") pod \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.375755 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-catalog-content\") pod \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\" (UID: \"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf\") " Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.376467 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-utilities" (OuterVolumeSpecName: "utilities") pod "21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" (UID: "21d73ca1-be43-4fb5-96d9-0e98f6a9bacf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.384653 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-kube-api-access-l7wdp" (OuterVolumeSpecName: "kube-api-access-l7wdp") pod "21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" (UID: "21d73ca1-be43-4fb5-96d9-0e98f6a9bacf"). InnerVolumeSpecName "kube-api-access-l7wdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.478536 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.478584 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7wdp\" (UniqueName: \"kubernetes.io/projected/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-kube-api-access-l7wdp\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.501005 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" (UID: "21d73ca1-be43-4fb5-96d9-0e98f6a9bacf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.580852 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.683773 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fg52v" event={"ID":"21d73ca1-be43-4fb5-96d9-0e98f6a9bacf","Type":"ContainerDied","Data":"2d8f82146887cccd5c6aa92c7e3562569d39ef9c946f4fa5ca58d4268d2c46a3"} Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.683827 4848 scope.go:117] "RemoveContainer" containerID="aff8a8d3d8f5f4d8e5f9a0e21bc65bdd09e5bbb65ce05a7cc3cfbc83f7f07e49" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.683964 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fg52v" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.712412 4848 scope.go:117] "RemoveContainer" containerID="e79615966c2c5527fbe43dfd762cd4bbb7df2e70cce109957ae826a93c316168" Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.721770 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fg52v"] Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.738607 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fg52v"] Dec 04 14:02:54 crc kubenswrapper[4848]: I1204 14:02:54.755467 4848 scope.go:117] "RemoveContainer" containerID="a142fa8f6fb5eec08276065d8446f0aa69622cae54846ae626c2b8816383677f" Dec 04 14:02:56 crc kubenswrapper[4848]: I1204 14:02:56.402497 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" path="/var/lib/kubelet/pods/21d73ca1-be43-4fb5-96d9-0e98f6a9bacf/volumes" Dec 04 14:03:06 crc kubenswrapper[4848]: I1204 14:03:06.827789 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-lgmcj" event={"ID":"5d3f5a36-9e68-4f3d-94b8-8cb59192211b","Type":"ContainerStarted","Data":"55b4f30740bd1f5758404d0131460f095f6ae6ce3f091fca4b3189ea645cfb7b"} Dec 04 14:03:06 crc kubenswrapper[4848]: I1204 14:03:06.831097 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" event={"ID":"461d1ece-26ff-40ed-a92c-f21300adeec7","Type":"ContainerStarted","Data":"baafc25f49097886308e9bd1cbd45f2c77010a58007e95c205f906cfefc269f1"} Dec 04 14:03:06 crc kubenswrapper[4848]: I1204 14:03:06.865242 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-lgmcj" podStartSLOduration=1.971704093 podStartE2EDuration="18.865224482s" podCreationTimestamp="2025-12-04 14:02:48 +0000 UTC" firstStartedPulling="2025-12-04 14:02:49.605073152 +0000 UTC m=+873.547569680" lastFinishedPulling="2025-12-04 14:03:06.498593551 +0000 UTC m=+890.441090069" observedRunningTime="2025-12-04 14:03:06.860669241 +0000 UTC m=+890.803165769" watchObservedRunningTime="2025-12-04 14:03:06.865224482 +0000 UTC m=+890.807721010" Dec 04 14:03:12 crc kubenswrapper[4848]: I1204 14:03:12.900660 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" event={"ID":"461d1ece-26ff-40ed-a92c-f21300adeec7","Type":"ContainerStarted","Data":"f8e9e4db690ad9fc7f26167c45d1a578883192d5e1c358a7bf81d9227785be79"} Dec 04 14:03:12 crc kubenswrapper[4848]: I1204 14:03:12.901155 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:03:12 crc kubenswrapper[4848]: I1204 14:03:12.903221 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" Dec 04 14:03:12 crc kubenswrapper[4848]: I1204 14:03:12.931273 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" podStartSLOduration=1.82509223 podStartE2EDuration="25.931251083s" podCreationTimestamp="2025-12-04 14:02:47 +0000 UTC" firstStartedPulling="2025-12-04 14:02:47.64738763 +0000 UTC m=+871.589884148" lastFinishedPulling="2025-12-04 14:03:11.753546473 +0000 UTC m=+895.696043001" observedRunningTime="2025-12-04 14:03:12.926155099 +0000 UTC m=+896.868651627" watchObservedRunningTime="2025-12-04 14:03:12.931251083 +0000 UTC m=+896.873747611" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.376547 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pk476"] Dec 04 14:03:15 crc kubenswrapper[4848]: E1204 14:03:15.377056 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="extract-utilities" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.377084 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="extract-utilities" Dec 04 14:03:15 crc kubenswrapper[4848]: E1204 14:03:15.377101 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="registry-server" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.377109 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="registry-server" Dec 04 14:03:15 crc kubenswrapper[4848]: E1204 14:03:15.377129 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="extract-content" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.377142 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="extract-content" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.377284 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d73ca1-be43-4fb5-96d9-0e98f6a9bacf" containerName="registry-server" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.378306 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.403568 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pk476"] Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.502633 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj2bn\" (UniqueName: \"kubernetes.io/projected/c6ef8633-f309-41f2-ad3c-59209be60d85-kube-api-access-fj2bn\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.502998 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-utilities\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.503653 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-catalog-content\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.605395 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj2bn\" (UniqueName: \"kubernetes.io/projected/c6ef8633-f309-41f2-ad3c-59209be60d85-kube-api-access-fj2bn\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.605469 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-utilities\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.605504 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-catalog-content\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.605926 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-catalog-content\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.606063 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-utilities\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.637073 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj2bn\" (UniqueName: \"kubernetes.io/projected/c6ef8633-f309-41f2-ad3c-59209be60d85-kube-api-access-fj2bn\") pod \"community-operators-pk476\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:15 crc kubenswrapper[4848]: I1204 14:03:15.693058 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:16 crc kubenswrapper[4848]: W1204 14:03:16.024027 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6ef8633_f309_41f2_ad3c_59209be60d85.slice/crio-aa5b6f9a161228a502278de761e4a7fed3e2f44f3d760df48f1865e9e346a408 WatchSource:0}: Error finding container aa5b6f9a161228a502278de761e4a7fed3e2f44f3d760df48f1865e9e346a408: Status 404 returned error can't find the container with id aa5b6f9a161228a502278de761e4a7fed3e2f44f3d760df48f1865e9e346a408 Dec 04 14:03:16 crc kubenswrapper[4848]: I1204 14:03:16.033813 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pk476"] Dec 04 14:03:16 crc kubenswrapper[4848]: I1204 14:03:16.926137 4848 generic.go:334] "Generic (PLEG): container finished" podID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerID="3bf487af9bb333cd052187274b7e7143778c876fd80539022a5e2269f5b7b66b" exitCode=0 Dec 04 14:03:16 crc kubenswrapper[4848]: I1204 14:03:16.926270 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk476" event={"ID":"c6ef8633-f309-41f2-ad3c-59209be60d85","Type":"ContainerDied","Data":"3bf487af9bb333cd052187274b7e7143778c876fd80539022a5e2269f5b7b66b"} Dec 04 14:03:16 crc kubenswrapper[4848]: I1204 14:03:16.926446 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk476" event={"ID":"c6ef8633-f309-41f2-ad3c-59209be60d85","Type":"ContainerStarted","Data":"aa5b6f9a161228a502278de761e4a7fed3e2f44f3d760df48f1865e9e346a408"} Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.669273 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.670384 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.672070 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.673137 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.673187 4848 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-bwr94" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.680023 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.838702 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e4f65a90-62e6-4183-ae02-ffa8bf65db2b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e4f65a90-62e6-4183-ae02-ffa8bf65db2b\") pod \"minio\" (UID: \"fa7f38d1-4bf1-4267-8f01-3ba86abcc903\") " pod="minio-dev/minio" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.838770 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z6qb\" (UniqueName: \"kubernetes.io/projected/fa7f38d1-4bf1-4267-8f01-3ba86abcc903-kube-api-access-2z6qb\") pod \"minio\" (UID: \"fa7f38d1-4bf1-4267-8f01-3ba86abcc903\") " pod="minio-dev/minio" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.935557 4848 generic.go:334] "Generic (PLEG): container finished" podID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerID="0c13583c94e04b420b358a3951f1ad76a94ec2e586eb3621e8fa22a695e1a09d" exitCode=0 Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.935596 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk476" event={"ID":"c6ef8633-f309-41f2-ad3c-59209be60d85","Type":"ContainerDied","Data":"0c13583c94e04b420b358a3951f1ad76a94ec2e586eb3621e8fa22a695e1a09d"} Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.940094 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e4f65a90-62e6-4183-ae02-ffa8bf65db2b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e4f65a90-62e6-4183-ae02-ffa8bf65db2b\") pod \"minio\" (UID: \"fa7f38d1-4bf1-4267-8f01-3ba86abcc903\") " pod="minio-dev/minio" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.940178 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z6qb\" (UniqueName: \"kubernetes.io/projected/fa7f38d1-4bf1-4267-8f01-3ba86abcc903-kube-api-access-2z6qb\") pod \"minio\" (UID: \"fa7f38d1-4bf1-4267-8f01-3ba86abcc903\") " pod="minio-dev/minio" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.942718 4848 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.942756 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e4f65a90-62e6-4183-ae02-ffa8bf65db2b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e4f65a90-62e6-4183-ae02-ffa8bf65db2b\") pod \"minio\" (UID: \"fa7f38d1-4bf1-4267-8f01-3ba86abcc903\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/72d082775282141670a9e911de5f204338eea3074d542070e90d435523c982ca/globalmount\"" pod="minio-dev/minio" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.962041 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z6qb\" (UniqueName: \"kubernetes.io/projected/fa7f38d1-4bf1-4267-8f01-3ba86abcc903-kube-api-access-2z6qb\") pod \"minio\" (UID: \"fa7f38d1-4bf1-4267-8f01-3ba86abcc903\") " pod="minio-dev/minio" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.980928 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e4f65a90-62e6-4183-ae02-ffa8bf65db2b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e4f65a90-62e6-4183-ae02-ffa8bf65db2b\") pod \"minio\" (UID: \"fa7f38d1-4bf1-4267-8f01-3ba86abcc903\") " pod="minio-dev/minio" Dec 04 14:03:17 crc kubenswrapper[4848]: I1204 14:03:17.991925 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 04 14:03:18 crc kubenswrapper[4848]: I1204 14:03:18.406002 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 04 14:03:18 crc kubenswrapper[4848]: W1204 14:03:18.414185 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa7f38d1_4bf1_4267_8f01_3ba86abcc903.slice/crio-c4f8d6fa79b8ae45a0da6bc480dac1ac619a7283a3601cfcc1ea473ca418f564 WatchSource:0}: Error finding container c4f8d6fa79b8ae45a0da6bc480dac1ac619a7283a3601cfcc1ea473ca418f564: Status 404 returned error can't find the container with id c4f8d6fa79b8ae45a0da6bc480dac1ac619a7283a3601cfcc1ea473ca418f564 Dec 04 14:03:18 crc kubenswrapper[4848]: I1204 14:03:18.943273 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"fa7f38d1-4bf1-4267-8f01-3ba86abcc903","Type":"ContainerStarted","Data":"c4f8d6fa79b8ae45a0da6bc480dac1ac619a7283a3601cfcc1ea473ca418f564"} Dec 04 14:03:18 crc kubenswrapper[4848]: I1204 14:03:18.948236 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk476" event={"ID":"c6ef8633-f309-41f2-ad3c-59209be60d85","Type":"ContainerStarted","Data":"d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087"} Dec 04 14:03:18 crc kubenswrapper[4848]: I1204 14:03:18.978340 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pk476" podStartSLOduration=2.272294396 podStartE2EDuration="3.97832201s" podCreationTimestamp="2025-12-04 14:03:15 +0000 UTC" firstStartedPulling="2025-12-04 14:03:16.927865656 +0000 UTC m=+900.870362194" lastFinishedPulling="2025-12-04 14:03:18.63389329 +0000 UTC m=+902.576389808" observedRunningTime="2025-12-04 14:03:18.971431762 +0000 UTC m=+902.913928320" watchObservedRunningTime="2025-12-04 14:03:18.97832201 +0000 UTC m=+902.920818538" Dec 04 14:03:21 crc kubenswrapper[4848]: I1204 14:03:21.969455 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"fa7f38d1-4bf1-4267-8f01-3ba86abcc903","Type":"ContainerStarted","Data":"b17911495dcdf7b29159ca17778c9cde9a07d19cc04b29464a163eaaf8d0570f"} Dec 04 14:03:21 crc kubenswrapper[4848]: I1204 14:03:21.993716 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.064552261 podStartE2EDuration="6.993694545s" podCreationTimestamp="2025-12-04 14:03:15 +0000 UTC" firstStartedPulling="2025-12-04 14:03:18.415668728 +0000 UTC m=+902.358165306" lastFinishedPulling="2025-12-04 14:03:21.344811062 +0000 UTC m=+905.287307590" observedRunningTime="2025-12-04 14:03:21.988304124 +0000 UTC m=+905.930800652" watchObservedRunningTime="2025-12-04 14:03:21.993694545 +0000 UTC m=+905.936191073" Dec 04 14:03:24 crc kubenswrapper[4848]: I1204 14:03:24.984268 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk"] Dec 04 14:03:24 crc kubenswrapper[4848]: I1204 14:03:24.985604 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:24 crc kubenswrapper[4848]: I1204 14:03:24.988020 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 04 14:03:24 crc kubenswrapper[4848]: I1204 14:03:24.989065 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 04 14:03:24 crc kubenswrapper[4848]: I1204 14:03:24.989362 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 04 14:03:24 crc kubenswrapper[4848]: I1204 14:03:24.990060 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-q2d7h" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.003630 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.015512 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.139861 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf9c3660-b2e2-4fac-a328-832144acf92c-config\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.140006 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.140058 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdh4b\" (UniqueName: \"kubernetes.io/projected/cf9c3660-b2e2-4fac-a328-832144acf92c-kube-api-access-sdh4b\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.140124 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.140207 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.156213 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-g78k5"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.157256 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.159124 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.163607 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.166873 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.180042 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-g78k5"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.232816 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.241359 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.242884 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdh4b\" (UniqueName: \"kubernetes.io/projected/cf9c3660-b2e2-4fac-a328-832144acf92c-kube-api-access-sdh4b\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.242929 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgsjz\" (UniqueName: \"kubernetes.io/projected/6d38276f-2908-496a-aa46-175984204c26-kube-api-access-cgsjz\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.242982 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.243008 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.243036 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d38276f-2908-496a-aa46-175984204c26-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.243059 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.243122 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d38276f-2908-496a-aa46-175984204c26-config\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.243156 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.243201 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.243241 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf9c3660-b2e2-4fac-a328-832144acf92c-config\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.243263 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.245281 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.245506 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.246579 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.247275 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf9c3660-b2e2-4fac-a328-832144acf92c-config\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.250483 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.255464 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/cf9c3660-b2e2-4fac-a328-832144acf92c-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.265160 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.268766 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdh4b\" (UniqueName: \"kubernetes.io/projected/cf9c3660-b2e2-4fac-a328-832144acf92c-kube-api-access-sdh4b\") pod \"logging-loki-distributor-76cc67bf56-mkvbk\" (UID: \"cf9c3660-b2e2-4fac-a328-832144acf92c\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.329719 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-7db4f94957-p9qcz"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.330021 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.330991 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.338676 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-7db4f94957-qqjgk"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.339713 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346701 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d38276f-2908-496a-aa46-175984204c26-config\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346745 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgs7n\" (UniqueName: \"kubernetes.io/projected/d659950d-e89c-45fa-90c5-4526593ed819-kube-api-access-sgs7n\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346774 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346806 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346829 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346845 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346886 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgsjz\" (UniqueName: \"kubernetes.io/projected/6d38276f-2908-496a-aa46-175984204c26-kube-api-access-cgsjz\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346905 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346920 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d659950d-e89c-45fa-90c5-4526593ed819-config\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346947 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d38276f-2908-496a-aa46-175984204c26-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.346975 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.347986 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d38276f-2908-496a-aa46-175984204c26-config\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.348445 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-njclc" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.349039 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.349131 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d38276f-2908-496a-aa46-175984204c26-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.349228 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.349293 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.349347 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.349470 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.350941 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.354474 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.363306 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-7db4f94957-p9qcz"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.369504 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/6d38276f-2908-496a-aa46-175984204c26-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.369873 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-7db4f94957-qqjgk"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.393641 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgsjz\" (UniqueName: \"kubernetes.io/projected/6d38276f-2908-496a-aa46-175984204c26-kube-api-access-cgsjz\") pod \"logging-loki-querier-5895d59bb8-g78k5\" (UID: \"6d38276f-2908-496a-aa46-175984204c26\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.448850 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgs7n\" (UniqueName: \"kubernetes.io/projected/d659950d-e89c-45fa-90c5-4526593ed819-kube-api-access-sgs7n\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.448911 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-lokistack-gateway\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.448982 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449015 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tls-secret\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449040 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgt5w\" (UniqueName: \"kubernetes.io/projected/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-kube-api-access-dgt5w\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449074 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-rbac\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449119 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449147 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449174 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449198 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4xlr\" (UniqueName: \"kubernetes.io/projected/3457940c-ae43-4e49-b616-a450d524ffd9-kube-api-access-f4xlr\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449371 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-rbac\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449452 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449540 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tls-secret\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449616 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d659950d-e89c-45fa-90c5-4526593ed819-config\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449635 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449664 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449711 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449753 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tenants\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449813 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tenants\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449835 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-lokistack-gateway\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.449871 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.450385 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.450443 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d659950d-e89c-45fa-90c5-4526593ed819-config\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.453586 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.453883 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/d659950d-e89c-45fa-90c5-4526593ed819-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.463662 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgs7n\" (UniqueName: \"kubernetes.io/projected/d659950d-e89c-45fa-90c5-4526593ed819-kube-api-access-sgs7n\") pod \"logging-loki-query-frontend-84558f7c9f-htkln\" (UID: \"d659950d-e89c-45fa-90c5-4526593ed819\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.479546 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.550926 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-rbac\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551264 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551286 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4xlr\" (UniqueName: \"kubernetes.io/projected/3457940c-ae43-4e49-b616-a450d524ffd9-kube-api-access-f4xlr\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551309 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-rbac\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551343 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551360 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tls-secret\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551388 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551406 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551431 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tenants\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551459 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tenants\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551477 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-lokistack-gateway\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551504 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551526 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-lokistack-gateway\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551563 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551580 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tls-secret\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551597 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgt5w\" (UniqueName: \"kubernetes.io/projected/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-kube-api-access-dgt5w\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.551768 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-rbac\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.556003 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: E1204 14:03:25.556496 4848 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Dec 04 14:03:25 crc kubenswrapper[4848]: E1204 14:03:25.556548 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tls-secret podName:bf4fb005-0de1-45b7-95a1-0fc7783e3a72 nodeName:}" failed. No retries permitted until 2025-12-04 14:03:26.056532308 +0000 UTC m=+909.999028836 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tls-secret") pod "logging-loki-gateway-7db4f94957-qqjgk" (UID: "bf4fb005-0de1-45b7-95a1-0fc7783e3a72") : secret "logging-loki-gateway-http" not found Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.556883 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-rbac\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.557354 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tenants\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: E1204 14:03:25.557430 4848 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Dec 04 14:03:25 crc kubenswrapper[4848]: E1204 14:03:25.557464 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tls-secret podName:3457940c-ae43-4e49-b616-a450d524ffd9 nodeName:}" failed. No retries permitted until 2025-12-04 14:03:26.05745315 +0000 UTC m=+909.999949768 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tls-secret") pod "logging-loki-gateway-7db4f94957-p9qcz" (UID: "3457940c-ae43-4e49-b616-a450d524ffd9") : secret "logging-loki-gateway-http" not found Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.557430 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.557644 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.558225 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/3457940c-ae43-4e49-b616-a450d524ffd9-lokistack-gateway\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.558274 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.558424 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-lokistack-gateway\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.564337 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tenants\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.565547 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.567634 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.575286 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgt5w\" (UniqueName: \"kubernetes.io/projected/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-kube-api-access-dgt5w\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.577122 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4xlr\" (UniqueName: \"kubernetes.io/projected/3457940c-ae43-4e49-b616-a450d524ffd9-kube-api-access-f4xlr\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.620622 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.649517 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk"] Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.694098 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.694222 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.760333 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:25 crc kubenswrapper[4848]: I1204 14:03:25.922988 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-g78k5"] Dec 04 14:03:25 crc kubenswrapper[4848]: W1204 14:03:25.926085 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d38276f_2908_496a_aa46_175984204c26.slice/crio-aa7697e8e04b39e16b00c5b8e3a46d6d2e1bbdce5ca552543aaf0eab3caf80cd WatchSource:0}: Error finding container aa7697e8e04b39e16b00c5b8e3a46d6d2e1bbdce5ca552543aaf0eab3caf80cd: Status 404 returned error can't find the container with id aa7697e8e04b39e16b00c5b8e3a46d6d2e1bbdce5ca552543aaf0eab3caf80cd Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.023479 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" event={"ID":"6d38276f-2908-496a-aa46-175984204c26","Type":"ContainerStarted","Data":"aa7697e8e04b39e16b00c5b8e3a46d6d2e1bbdce5ca552543aaf0eab3caf80cd"} Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.025345 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" event={"ID":"cf9c3660-b2e2-4fac-a328-832144acf92c","Type":"ContainerStarted","Data":"e3290a3487e2efaa31898cf8b017bba84524940ec83a376034d64237d02ab3b0"} Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.061763 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tls-secret\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.061893 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tls-secret\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.066322 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bf4fb005-0de1-45b7-95a1-0fc7783e3a72-tls-secret\") pod \"logging-loki-gateway-7db4f94957-qqjgk\" (UID: \"bf4fb005-0de1-45b7-95a1-0fc7783e3a72\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.066687 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/3457940c-ae43-4e49-b616-a450d524ffd9-tls-secret\") pod \"logging-loki-gateway-7db4f94957-p9qcz\" (UID: \"3457940c-ae43-4e49-b616-a450d524ffd9\") " pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.092826 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.098846 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.140371 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.141247 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.143277 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.143509 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.156453 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pk476"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.162337 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.234152 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.235484 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.239881 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.240221 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.242635 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.264235 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-36b11453-ee3d-475f-9cf3-1e8152ac4077\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-36b11453-ee3d-475f-9cf3-1e8152ac4077\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.264299 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.264329 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.264366 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47jkd\" (UniqueName: \"kubernetes.io/projected/19d88d7c-8641-499e-8fc1-8864af7feb3c-kube-api-access-47jkd\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.264400 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.264421 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d88d7c-8641-499e-8fc1-8864af7feb3c-config\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.264439 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-df24c8a8-ffe6-41d8-9c1d-da236de9e62b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-df24c8a8-ffe6-41d8-9c1d-da236de9e62b\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.264468 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.310939 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.312706 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.314658 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.314870 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.317174 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.318454 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.331873 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365379 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47jkd\" (UniqueName: \"kubernetes.io/projected/19d88d7c-8641-499e-8fc1-8864af7feb3c-kube-api-access-47jkd\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365440 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365463 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365484 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-config\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365504 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d88d7c-8641-499e-8fc1-8864af7feb3c-config\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365526 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-df24c8a8-ffe6-41d8-9c1d-da236de9e62b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-df24c8a8-ffe6-41d8-9c1d-da236de9e62b\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365556 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365582 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365605 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365623 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjvd6\" (UniqueName: \"kubernetes.io/projected/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-kube-api-access-hjvd6\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365659 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-36b11453-ee3d-475f-9cf3-1e8152ac4077\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-36b11453-ee3d-475f-9cf3-1e8152ac4077\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365677 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365700 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365721 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-163c9f65-eafa-4684-82f5-416916d6fd51\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-163c9f65-eafa-4684-82f5-416916d6fd51\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.365743 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.367796 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.368495 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d88d7c-8641-499e-8fc1-8864af7feb3c-config\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.370345 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.370963 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.371203 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/19d88d7c-8641-499e-8fc1-8864af7feb3c-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.371697 4848 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.371724 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-36b11453-ee3d-475f-9cf3-1e8152ac4077\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-36b11453-ee3d-475f-9cf3-1e8152ac4077\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/25338c6a7912bd72cbca336ec219f4a7e5561ffb0fafa9505e6530d6f22ba673/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.383228 4848 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.383267 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-df24c8a8-ffe6-41d8-9c1d-da236de9e62b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-df24c8a8-ffe6-41d8-9c1d-da236de9e62b\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/35a354a1e5dd2f790585d6f9ece67ce88e626c10e8d4b0d4d3bf5c239c461a59/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.389648 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47jkd\" (UniqueName: \"kubernetes.io/projected/19d88d7c-8641-499e-8fc1-8864af7feb3c-kube-api-access-47jkd\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.413162 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-df24c8a8-ffe6-41d8-9c1d-da236de9e62b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-df24c8a8-ffe6-41d8-9c1d-da236de9e62b\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.420443 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-36b11453-ee3d-475f-9cf3-1e8152ac4077\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-36b11453-ee3d-475f-9cf3-1e8152ac4077\") pod \"logging-loki-ingester-0\" (UID: \"19d88d7c-8641-499e-8fc1-8864af7feb3c\") " pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.465665 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467035 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25f82b9f-ffd8-47b2-9abe-593234820cb5-config\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467096 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467145 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-config\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467164 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467187 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjvd6\" (UniqueName: \"kubernetes.io/projected/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-kube-api-access-hjvd6\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467219 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467250 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-46898e86-0df5-49d4-be8c-d4d1280d649a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-46898e86-0df5-49d4-be8c-d4d1280d649a\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467286 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467305 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-163c9f65-eafa-4684-82f5-416916d6fd51\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-163c9f65-eafa-4684-82f5-416916d6fd51\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467328 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467362 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg9bz\" (UniqueName: \"kubernetes.io/projected/25f82b9f-ffd8-47b2-9abe-593234820cb5-kube-api-access-vg9bz\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467402 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467423 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.467449 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.470462 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-config\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.470742 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.474503 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.474904 4848 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.474933 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-163c9f65-eafa-4684-82f5-416916d6fd51\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-163c9f65-eafa-4684-82f5-416916d6fd51\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/94fd19d3cc82236f60048a5e5b5d66f81d68e992caeb2cf1cef50e881110a52f/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.483432 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.484575 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.485243 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjvd6\" (UniqueName: \"kubernetes.io/projected/b53d48d7-e9d7-442d-b2b8-044a7eb9e692-kube-api-access-hjvd6\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.506870 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-163c9f65-eafa-4684-82f5-416916d6fd51\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-163c9f65-eafa-4684-82f5-416916d6fd51\") pod \"logging-loki-compactor-0\" (UID: \"b53d48d7-e9d7-442d-b2b8-044a7eb9e692\") " pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.569157 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.569248 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.569277 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-46898e86-0df5-49d4-be8c-d4d1280d649a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-46898e86-0df5-49d4-be8c-d4d1280d649a\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.569312 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.569335 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg9bz\" (UniqueName: \"kubernetes.io/projected/25f82b9f-ffd8-47b2-9abe-593234820cb5-kube-api-access-vg9bz\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.569378 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.569444 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25f82b9f-ffd8-47b2-9abe-593234820cb5-config\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.570479 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25f82b9f-ffd8-47b2-9abe-593234820cb5-config\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.571130 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.574010 4848 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.574053 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-46898e86-0df5-49d4-be8c-d4d1280d649a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-46898e86-0df5-49d4-be8c-d4d1280d649a\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2e5ff5c0edc4e067cd25224fa2b786b7eab2589b8f0054527468ff69b8dd0ab0/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.574297 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.574452 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.574477 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/25f82b9f-ffd8-47b2-9abe-593234820cb5-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.588435 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg9bz\" (UniqueName: \"kubernetes.io/projected/25f82b9f-ffd8-47b2-9abe-593234820cb5-kube-api-access-vg9bz\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.613046 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-46898e86-0df5-49d4-be8c-d4d1280d649a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-46898e86-0df5-49d4-be8c-d4d1280d649a\") pod \"logging-loki-index-gateway-0\" (UID: \"25f82b9f-ffd8-47b2-9abe-593234820cb5\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.635843 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.748801 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-7db4f94957-p9qcz"] Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.774816 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.818431 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-7db4f94957-qqjgk"] Dec 04 14:03:26 crc kubenswrapper[4848]: W1204 14:03:26.821159 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf4fb005_0de1_45b7_95a1_0fc7783e3a72.slice/crio-3065b3c7d6570943a4eecf93176209e5dbb94a65e564e2e6dec39e2a983df737 WatchSource:0}: Error finding container 3065b3c7d6570943a4eecf93176209e5dbb94a65e564e2e6dec39e2a983df737: Status 404 returned error can't find the container with id 3065b3c7d6570943a4eecf93176209e5dbb94a65e564e2e6dec39e2a983df737 Dec 04 14:03:26 crc kubenswrapper[4848]: I1204 14:03:26.903100 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 04 14:03:27 crc kubenswrapper[4848]: I1204 14:03:27.033123 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" event={"ID":"d659950d-e89c-45fa-90c5-4526593ed819","Type":"ContainerStarted","Data":"279490bfc46e33b550cb132672aca40c0fd21b5afb71b6ccb1d822c1f3acc4c7"} Dec 04 14:03:27 crc kubenswrapper[4848]: I1204 14:03:27.034359 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" event={"ID":"3457940c-ae43-4e49-b616-a450d524ffd9","Type":"ContainerStarted","Data":"774cb5e636535b41d8032d76428e792f960223f2b43c169c2a7d3049c2b4f38a"} Dec 04 14:03:27 crc kubenswrapper[4848]: I1204 14:03:27.035325 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" event={"ID":"bf4fb005-0de1-45b7-95a1-0fc7783e3a72","Type":"ContainerStarted","Data":"3065b3c7d6570943a4eecf93176209e5dbb94a65e564e2e6dec39e2a983df737"} Dec 04 14:03:27 crc kubenswrapper[4848]: I1204 14:03:27.036119 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"19d88d7c-8641-499e-8fc1-8864af7feb3c","Type":"ContainerStarted","Data":"df1cfe8c0775ee4509127f9457c3560e2b5015b8ee430b9e7f1bc1859cdc06b7"} Dec 04 14:03:27 crc kubenswrapper[4848]: I1204 14:03:27.056173 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 04 14:03:27 crc kubenswrapper[4848]: I1204 14:03:27.205634 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 04 14:03:27 crc kubenswrapper[4848]: W1204 14:03:27.222544 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25f82b9f_ffd8_47b2_9abe_593234820cb5.slice/crio-5e3a705daa18ef218202d9270b7c393d4abc9b573000cb003fb5e6e1b3388545 WatchSource:0}: Error finding container 5e3a705daa18ef218202d9270b7c393d4abc9b573000cb003fb5e6e1b3388545: Status 404 returned error can't find the container with id 5e3a705daa18ef218202d9270b7c393d4abc9b573000cb003fb5e6e1b3388545 Dec 04 14:03:28 crc kubenswrapper[4848]: I1204 14:03:28.058665 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"25f82b9f-ffd8-47b2-9abe-593234820cb5","Type":"ContainerStarted","Data":"5e3a705daa18ef218202d9270b7c393d4abc9b573000cb003fb5e6e1b3388545"} Dec 04 14:03:28 crc kubenswrapper[4848]: I1204 14:03:28.060909 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"b53d48d7-e9d7-442d-b2b8-044a7eb9e692","Type":"ContainerStarted","Data":"5d3434a984f9e8f93eea146a53d85a322464336f8385745b39e86d9b6169cafd"} Dec 04 14:03:28 crc kubenswrapper[4848]: I1204 14:03:28.061073 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pk476" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="registry-server" containerID="cri-o://d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087" gracePeriod=2 Dec 04 14:03:30 crc kubenswrapper[4848]: I1204 14:03:30.078447 4848 generic.go:334] "Generic (PLEG): container finished" podID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerID="d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087" exitCode=0 Dec 04 14:03:30 crc kubenswrapper[4848]: I1204 14:03:30.078496 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk476" event={"ID":"c6ef8633-f309-41f2-ad3c-59209be60d85","Type":"ContainerDied","Data":"d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087"} Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.533155 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hbp7n"] Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.536488 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.548402 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hbp7n"] Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.590453 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-catalog-content\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.590548 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-utilities\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.590673 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clznh\" (UniqueName: \"kubernetes.io/projected/aeb40a68-89aa-4358-9912-4e87ec546a83-kube-api-access-clznh\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.692572 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-catalog-content\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.692688 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-utilities\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.692716 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clznh\" (UniqueName: \"kubernetes.io/projected/aeb40a68-89aa-4358-9912-4e87ec546a83-kube-api-access-clznh\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.693242 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-catalog-content\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.693595 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-utilities\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.718332 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clznh\" (UniqueName: \"kubernetes.io/projected/aeb40a68-89aa-4358-9912-4e87ec546a83-kube-api-access-clznh\") pod \"certified-operators-hbp7n\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:33 crc kubenswrapper[4848]: I1204 14:03:33.861398 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:35 crc kubenswrapper[4848]: E1204 14:03:35.696469 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087 is running failed: container process not found" containerID="d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:03:35 crc kubenswrapper[4848]: E1204 14:03:35.698005 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087 is running failed: container process not found" containerID="d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:03:35 crc kubenswrapper[4848]: E1204 14:03:35.698382 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087 is running failed: container process not found" containerID="d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:03:35 crc kubenswrapper[4848]: E1204 14:03:35.698443 4848 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-pk476" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="registry-server" Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.129373 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.137724 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk476" event={"ID":"c6ef8633-f309-41f2-ad3c-59209be60d85","Type":"ContainerDied","Data":"aa5b6f9a161228a502278de761e4a7fed3e2f44f3d760df48f1865e9e346a408"} Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.137778 4848 scope.go:117] "RemoveContainer" containerID="d903b5df3d239165274beb6b5db84f86f9ef6ce223892762e4b363e542d50087" Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.309535 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-utilities\") pod \"c6ef8633-f309-41f2-ad3c-59209be60d85\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.309716 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj2bn\" (UniqueName: \"kubernetes.io/projected/c6ef8633-f309-41f2-ad3c-59209be60d85-kube-api-access-fj2bn\") pod \"c6ef8633-f309-41f2-ad3c-59209be60d85\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.310322 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-utilities" (OuterVolumeSpecName: "utilities") pod "c6ef8633-f309-41f2-ad3c-59209be60d85" (UID: "c6ef8633-f309-41f2-ad3c-59209be60d85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.310684 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-catalog-content\") pod \"c6ef8633-f309-41f2-ad3c-59209be60d85\" (UID: \"c6ef8633-f309-41f2-ad3c-59209be60d85\") " Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.311174 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.325287 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6ef8633-f309-41f2-ad3c-59209be60d85-kube-api-access-fj2bn" (OuterVolumeSpecName: "kube-api-access-fj2bn") pod "c6ef8633-f309-41f2-ad3c-59209be60d85" (UID: "c6ef8633-f309-41f2-ad3c-59209be60d85"). InnerVolumeSpecName "kube-api-access-fj2bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.374101 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6ef8633-f309-41f2-ad3c-59209be60d85" (UID: "c6ef8633-f309-41f2-ad3c-59209be60d85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.412616 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6ef8633-f309-41f2-ad3c-59209be60d85-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:36 crc kubenswrapper[4848]: I1204 14:03:36.412648 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj2bn\" (UniqueName: \"kubernetes.io/projected/c6ef8633-f309-41f2-ad3c-59209be60d85-kube-api-access-fj2bn\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:37 crc kubenswrapper[4848]: I1204 14:03:37.055702 4848 scope.go:117] "RemoveContainer" containerID="0c13583c94e04b420b358a3951f1ad76a94ec2e586eb3621e8fa22a695e1a09d" Dec 04 14:03:37 crc kubenswrapper[4848]: I1204 14:03:37.119812 4848 scope.go:117] "RemoveContainer" containerID="3bf487af9bb333cd052187274b7e7143778c876fd80539022a5e2269f5b7b66b" Dec 04 14:03:37 crc kubenswrapper[4848]: I1204 14:03:37.148854 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk476" Dec 04 14:03:37 crc kubenswrapper[4848]: I1204 14:03:37.174993 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pk476"] Dec 04 14:03:37 crc kubenswrapper[4848]: I1204 14:03:37.189992 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pk476"] Dec 04 14:03:37 crc kubenswrapper[4848]: I1204 14:03:37.539171 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hbp7n"] Dec 04 14:03:37 crc kubenswrapper[4848]: W1204 14:03:37.550583 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaeb40a68_89aa_4358_9912_4e87ec546a83.slice/crio-4ddc39e2183fcabf53aa3bf5afdc7ae61444d1957f57c8eb9e1e6da38eeeb6c0 WatchSource:0}: Error finding container 4ddc39e2183fcabf53aa3bf5afdc7ae61444d1957f57c8eb9e1e6da38eeeb6c0: Status 404 returned error can't find the container with id 4ddc39e2183fcabf53aa3bf5afdc7ae61444d1957f57c8eb9e1e6da38eeeb6c0 Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.159160 4848 generic.go:334] "Generic (PLEG): container finished" podID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerID="ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61" exitCode=0 Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.159274 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbp7n" event={"ID":"aeb40a68-89aa-4358-9912-4e87ec546a83","Type":"ContainerDied","Data":"ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.160895 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbp7n" event={"ID":"aeb40a68-89aa-4358-9912-4e87ec546a83","Type":"ContainerStarted","Data":"4ddc39e2183fcabf53aa3bf5afdc7ae61444d1957f57c8eb9e1e6da38eeeb6c0"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.168265 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" event={"ID":"cf9c3660-b2e2-4fac-a328-832144acf92c","Type":"ContainerStarted","Data":"92ee93e4e56c45e0879d16f89943c2b03c157d6c1ddec04389aa93291ff109b6"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.168608 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.170124 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" event={"ID":"bf4fb005-0de1-45b7-95a1-0fc7783e3a72","Type":"ContainerStarted","Data":"0711e090337e8c869a1275cb7fb172f22d880c138be99b532ffb9bb548fa4adb"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.171693 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"b53d48d7-e9d7-442d-b2b8-044a7eb9e692","Type":"ContainerStarted","Data":"dc5ab37fa9d538fbf0d0735ad4b824da8d0dc365304b03eb9f89166f35294dfb"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.172009 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.173101 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"25f82b9f-ffd8-47b2-9abe-593234820cb5","Type":"ContainerStarted","Data":"49951efff41bf42b20cfb97d460d5538ec9bb15df4f407deef534df961bec840"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.173241 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.174578 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" event={"ID":"d659950d-e89c-45fa-90c5-4526593ed819","Type":"ContainerStarted","Data":"5348f94ba878cacb6c67381ae7d68bfbed331817da1de04ca6ff4a17932f468e"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.174880 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.176110 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" event={"ID":"3457940c-ae43-4e49-b616-a450d524ffd9","Type":"ContainerStarted","Data":"9061c7387c8743290452a0052fe1c2c406633dcfd658cdea1f084de7c1ab7fe8"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.177939 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"19d88d7c-8641-499e-8fc1-8864af7feb3c","Type":"ContainerStarted","Data":"69d890b9b686c51de3648b5263b8fe8303b12719a535a5a128bb6d10f961d3a1"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.178741 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.180315 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" event={"ID":"6d38276f-2908-496a-aa46-175984204c26","Type":"ContainerStarted","Data":"5a000f0bada811072168b6d2e66309ae2e05d13c80d65f37dca75e1fc9a69e4e"} Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.180497 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.228786 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=2.9038872380000003 podStartE2EDuration="13.22876422s" podCreationTimestamp="2025-12-04 14:03:25 +0000 UTC" firstStartedPulling="2025-12-04 14:03:26.919385122 +0000 UTC m=+910.861881650" lastFinishedPulling="2025-12-04 14:03:37.244262094 +0000 UTC m=+921.186758632" observedRunningTime="2025-12-04 14:03:38.218413668 +0000 UTC m=+922.160910226" watchObservedRunningTime="2025-12-04 14:03:38.22876422 +0000 UTC m=+922.171260758" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.258304 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.236253715 podStartE2EDuration="13.258285109s" podCreationTimestamp="2025-12-04 14:03:25 +0000 UTC" firstStartedPulling="2025-12-04 14:03:27.224851954 +0000 UTC m=+911.167348482" lastFinishedPulling="2025-12-04 14:03:37.246883338 +0000 UTC m=+921.189379876" observedRunningTime="2025-12-04 14:03:38.2423073 +0000 UTC m=+922.184803828" watchObservedRunningTime="2025-12-04 14:03:38.258285109 +0000 UTC m=+922.200781637" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.271621 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" podStartSLOduration=2.062319994 podStartE2EDuration="13.271597074s" podCreationTimestamp="2025-12-04 14:03:25 +0000 UTC" firstStartedPulling="2025-12-04 14:03:25.929235928 +0000 UTC m=+909.871732456" lastFinishedPulling="2025-12-04 14:03:37.138513008 +0000 UTC m=+921.081009536" observedRunningTime="2025-12-04 14:03:38.262582214 +0000 UTC m=+922.205078782" watchObservedRunningTime="2025-12-04 14:03:38.271597074 +0000 UTC m=+922.214093612" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.291327 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" podStartSLOduration=2.144348973 podStartE2EDuration="13.291304244s" podCreationTimestamp="2025-12-04 14:03:25 +0000 UTC" firstStartedPulling="2025-12-04 14:03:26.099942537 +0000 UTC m=+910.042439065" lastFinishedPulling="2025-12-04 14:03:37.246897798 +0000 UTC m=+921.189394336" observedRunningTime="2025-12-04 14:03:38.282344966 +0000 UTC m=+922.224841504" watchObservedRunningTime="2025-12-04 14:03:38.291304244 +0000 UTC m=+922.233800782" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.304996 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.117155163 podStartE2EDuration="13.304977448s" podCreationTimestamp="2025-12-04 14:03:25 +0000 UTC" firstStartedPulling="2025-12-04 14:03:27.058653034 +0000 UTC m=+911.001149562" lastFinishedPulling="2025-12-04 14:03:37.246475299 +0000 UTC m=+921.188971847" observedRunningTime="2025-12-04 14:03:38.30016935 +0000 UTC m=+922.242665918" watchObservedRunningTime="2025-12-04 14:03:38.304977448 +0000 UTC m=+922.247473986" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.325325 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" podStartSLOduration=2.810794756 podStartE2EDuration="14.325305912s" podCreationTimestamp="2025-12-04 14:03:24 +0000 UTC" firstStartedPulling="2025-12-04 14:03:25.658669246 +0000 UTC m=+909.601165774" lastFinishedPulling="2025-12-04 14:03:37.173180402 +0000 UTC m=+921.115676930" observedRunningTime="2025-12-04 14:03:38.319630524 +0000 UTC m=+922.262127092" watchObservedRunningTime="2025-12-04 14:03:38.325305912 +0000 UTC m=+922.267802440" Dec 04 14:03:38 crc kubenswrapper[4848]: I1204 14:03:38.405637 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" path="/var/lib/kubelet/pods/c6ef8633-f309-41f2-ad3c-59209be60d85/volumes" Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.197539 4848 generic.go:334] "Generic (PLEG): container finished" podID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerID="df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264" exitCode=0 Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.197590 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbp7n" event={"ID":"aeb40a68-89aa-4358-9912-4e87ec546a83","Type":"ContainerDied","Data":"df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264"} Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.199909 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" event={"ID":"3457940c-ae43-4e49-b616-a450d524ffd9","Type":"ContainerStarted","Data":"7cbbc7417bbdfd8d8fe0602080668b4a8f378556f0176e3a0ba75cac4ffaf390"} Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.200058 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.200084 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.202117 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" event={"ID":"bf4fb005-0de1-45b7-95a1-0fc7783e3a72","Type":"ContainerStarted","Data":"663d6bd60b4e6d4c9b430d12028112b7d1427e488f987707e189d19c2263d435"} Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.202343 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.211737 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.213155 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.220159 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.244239 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" podStartSLOduration=2.499556078 podStartE2EDuration="15.244213915s" podCreationTimestamp="2025-12-04 14:03:25 +0000 UTC" firstStartedPulling="2025-12-04 14:03:26.823706111 +0000 UTC m=+910.766202639" lastFinishedPulling="2025-12-04 14:03:39.568363948 +0000 UTC m=+923.510860476" observedRunningTime="2025-12-04 14:03:40.237837939 +0000 UTC m=+924.180334467" watchObservedRunningTime="2025-12-04 14:03:40.244213915 +0000 UTC m=+924.186710463" Dec 04 14:03:40 crc kubenswrapper[4848]: I1204 14:03:40.263455 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-7db4f94957-p9qcz" podStartSLOduration=2.462588646 podStartE2EDuration="15.263430172s" podCreationTimestamp="2025-12-04 14:03:25 +0000 UTC" firstStartedPulling="2025-12-04 14:03:26.764240182 +0000 UTC m=+910.706736710" lastFinishedPulling="2025-12-04 14:03:39.565081708 +0000 UTC m=+923.507578236" observedRunningTime="2025-12-04 14:03:40.257850906 +0000 UTC m=+924.200347424" watchObservedRunningTime="2025-12-04 14:03:40.263430172 +0000 UTC m=+924.205926710" Dec 04 14:03:41 crc kubenswrapper[4848]: I1204 14:03:41.212378 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbp7n" event={"ID":"aeb40a68-89aa-4358-9912-4e87ec546a83","Type":"ContainerStarted","Data":"8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909"} Dec 04 14:03:41 crc kubenswrapper[4848]: I1204 14:03:41.213398 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:41 crc kubenswrapper[4848]: I1204 14:03:41.224935 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-7db4f94957-qqjgk" Dec 04 14:03:41 crc kubenswrapper[4848]: I1204 14:03:41.231418 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hbp7n" podStartSLOduration=5.714940686 podStartE2EDuration="8.231399856s" podCreationTimestamp="2025-12-04 14:03:33 +0000 UTC" firstStartedPulling="2025-12-04 14:03:38.162151328 +0000 UTC m=+922.104647846" lastFinishedPulling="2025-12-04 14:03:40.678610488 +0000 UTC m=+924.621107016" observedRunningTime="2025-12-04 14:03:41.229784267 +0000 UTC m=+925.172280795" watchObservedRunningTime="2025-12-04 14:03:41.231399856 +0000 UTC m=+925.173896384" Dec 04 14:03:43 crc kubenswrapper[4848]: I1204 14:03:43.862166 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:43 crc kubenswrapper[4848]: I1204 14:03:43.862630 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:43 crc kubenswrapper[4848]: I1204 14:03:43.950518 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:45 crc kubenswrapper[4848]: I1204 14:03:45.285693 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:45 crc kubenswrapper[4848]: I1204 14:03:45.352662 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hbp7n"] Dec 04 14:03:47 crc kubenswrapper[4848]: I1204 14:03:47.261540 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hbp7n" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerName="registry-server" containerID="cri-o://8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909" gracePeriod=2 Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.164675 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.209698 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-catalog-content\") pod \"aeb40a68-89aa-4358-9912-4e87ec546a83\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.209775 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-utilities\") pod \"aeb40a68-89aa-4358-9912-4e87ec546a83\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.209803 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clznh\" (UniqueName: \"kubernetes.io/projected/aeb40a68-89aa-4358-9912-4e87ec546a83-kube-api-access-clznh\") pod \"aeb40a68-89aa-4358-9912-4e87ec546a83\" (UID: \"aeb40a68-89aa-4358-9912-4e87ec546a83\") " Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.211374 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-utilities" (OuterVolumeSpecName: "utilities") pod "aeb40a68-89aa-4358-9912-4e87ec546a83" (UID: "aeb40a68-89aa-4358-9912-4e87ec546a83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.217285 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb40a68-89aa-4358-9912-4e87ec546a83-kube-api-access-clznh" (OuterVolumeSpecName: "kube-api-access-clznh") pod "aeb40a68-89aa-4358-9912-4e87ec546a83" (UID: "aeb40a68-89aa-4358-9912-4e87ec546a83"). InnerVolumeSpecName "kube-api-access-clznh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.265922 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aeb40a68-89aa-4358-9912-4e87ec546a83" (UID: "aeb40a68-89aa-4358-9912-4e87ec546a83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.270421 4848 generic.go:334] "Generic (PLEG): container finished" podID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerID="8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909" exitCode=0 Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.270463 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbp7n" event={"ID":"aeb40a68-89aa-4358-9912-4e87ec546a83","Type":"ContainerDied","Data":"8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909"} Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.270514 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbp7n" event={"ID":"aeb40a68-89aa-4358-9912-4e87ec546a83","Type":"ContainerDied","Data":"4ddc39e2183fcabf53aa3bf5afdc7ae61444d1957f57c8eb9e1e6da38eeeb6c0"} Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.270509 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbp7n" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.270530 4848 scope.go:117] "RemoveContainer" containerID="8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.304813 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hbp7n"] Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.305119 4848 scope.go:117] "RemoveContainer" containerID="df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.309484 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hbp7n"] Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.311478 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.311502 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeb40a68-89aa-4358-9912-4e87ec546a83-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.311512 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clznh\" (UniqueName: \"kubernetes.io/projected/aeb40a68-89aa-4358-9912-4e87ec546a83-kube-api-access-clznh\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.330620 4848 scope.go:117] "RemoveContainer" containerID="ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.360992 4848 scope.go:117] "RemoveContainer" containerID="8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909" Dec 04 14:03:48 crc kubenswrapper[4848]: E1204 14:03:48.361673 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909\": container with ID starting with 8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909 not found: ID does not exist" containerID="8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.361720 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909"} err="failed to get container status \"8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909\": rpc error: code = NotFound desc = could not find container \"8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909\": container with ID starting with 8c952f6392ba6695de4e432f35a84631db66a7b7a5e9b6475bc7b586711a1909 not found: ID does not exist" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.361747 4848 scope.go:117] "RemoveContainer" containerID="df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264" Dec 04 14:03:48 crc kubenswrapper[4848]: E1204 14:03:48.362189 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264\": container with ID starting with df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264 not found: ID does not exist" containerID="df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.362234 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264"} err="failed to get container status \"df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264\": rpc error: code = NotFound desc = could not find container \"df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264\": container with ID starting with df8d2bddb94dc30e09741c215a83e86e960512390b6886544007098097a57264 not found: ID does not exist" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.362265 4848 scope.go:117] "RemoveContainer" containerID="ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61" Dec 04 14:03:48 crc kubenswrapper[4848]: E1204 14:03:48.362752 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61\": container with ID starting with ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61 not found: ID does not exist" containerID="ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.362787 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61"} err="failed to get container status \"ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61\": rpc error: code = NotFound desc = could not find container \"ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61\": container with ID starting with ac154e9ac68e49d3c38b6f48c4ea2b4a6edc2fb8a70b5e8f67bfad02f9f81e61 not found: ID does not exist" Dec 04 14:03:48 crc kubenswrapper[4848]: I1204 14:03:48.413020 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" path="/var/lib/kubelet/pods/aeb40a68-89aa-4358-9912-4e87ec546a83/volumes" Dec 04 14:03:55 crc kubenswrapper[4848]: I1204 14:03:55.336721 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-mkvbk" Dec 04 14:03:55 crc kubenswrapper[4848]: I1204 14:03:55.524260 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-g78k5" Dec 04 14:03:55 crc kubenswrapper[4848]: I1204 14:03:55.654518 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-htkln" Dec 04 14:03:56 crc kubenswrapper[4848]: I1204 14:03:56.476750 4848 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 04 14:03:56 crc kubenswrapper[4848]: I1204 14:03:56.477072 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="19d88d7c-8641-499e-8fc1-8864af7feb3c" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 04 14:03:56 crc kubenswrapper[4848]: I1204 14:03:56.645468 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 04 14:03:56 crc kubenswrapper[4848]: I1204 14:03:56.785710 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 04 14:04:06 crc kubenswrapper[4848]: I1204 14:04:06.474997 4848 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 04 14:04:06 crc kubenswrapper[4848]: I1204 14:04:06.475661 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="19d88d7c-8641-499e-8fc1-8864af7feb3c" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 04 14:04:14 crc kubenswrapper[4848]: I1204 14:04:14.314494 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:04:14 crc kubenswrapper[4848]: I1204 14:04:14.314912 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:04:16 crc kubenswrapper[4848]: I1204 14:04:16.481098 4848 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 04 14:04:16 crc kubenswrapper[4848]: I1204 14:04:16.481438 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="19d88d7c-8641-499e-8fc1-8864af7feb3c" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 04 14:04:26 crc kubenswrapper[4848]: I1204 14:04:26.470770 4848 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 04 14:04:26 crc kubenswrapper[4848]: I1204 14:04:26.471159 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="19d88d7c-8641-499e-8fc1-8864af7feb3c" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 04 14:04:36 crc kubenswrapper[4848]: I1204 14:04:36.470569 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 04 14:04:44 crc kubenswrapper[4848]: I1204 14:04:44.314083 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:04:44 crc kubenswrapper[4848]: I1204 14:04:44.314915 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.030273 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-5v5f2"] Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.032301 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerName="extract-utilities" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.032459 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerName="extract-utilities" Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.032611 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerName="registry-server" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.032717 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerName="registry-server" Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.032848 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="extract-utilities" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.033008 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="extract-utilities" Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.033156 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="registry-server" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.033268 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="registry-server" Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.033388 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerName="extract-content" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.033527 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerName="extract-content" Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.033657 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="extract-content" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.033780 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="extract-content" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.034172 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6ef8633-f309-41f2-ad3c-59209be60d85" containerName="registry-server" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.034315 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb40a68-89aa-4358-9912-4e87ec546a83" containerName="registry-server" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.035266 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.038138 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-7nz9k" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.038912 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.039172 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.039213 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.040691 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.063136 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-5v5f2"] Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.082446 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131557 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131634 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config-openshift-service-cacrt\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131684 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64ffd\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-kube-api-access-64ffd\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131717 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-entrypoint\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131734 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-datadir\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131754 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131769 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-trusted-ca\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131813 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-tmp\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131843 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-token\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131861 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-sa-token\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.131879 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.191307 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-5v5f2"] Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.192167 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-64ffd metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-5v5f2" podUID="e981a4b5-8c6c-4d7e-a342-8972a8c69cee" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233455 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-token\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233499 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-sa-token\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233523 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233562 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233596 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config-openshift-service-cacrt\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233621 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64ffd\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-kube-api-access-64ffd\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233651 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-entrypoint\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233669 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-datadir\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233683 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.233686 4848 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.233704 4848 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233696 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-trusted-ca\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.233773 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics podName:e981a4b5-8c6c-4d7e-a342-8972a8c69cee nodeName:}" failed. No retries permitted until 2025-12-04 14:04:47.733751517 +0000 UTC m=+991.676248045 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics") pod "collector-5v5f2" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee") : secret "collector-metrics" not found Dec 04 14:04:47 crc kubenswrapper[4848]: E1204 14:04:47.233809 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver podName:e981a4b5-8c6c-4d7e-a342-8972a8c69cee nodeName:}" failed. No retries permitted until 2025-12-04 14:04:47.733798949 +0000 UTC m=+991.676295587 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver") pod "collector-5v5f2" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee") : secret "collector-syslog-receiver" not found Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.233829 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-tmp\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.234104 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-datadir\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.234584 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-entrypoint\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.234795 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.235413 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-trusted-ca\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.235442 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config-openshift-service-cacrt\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.246330 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-tmp\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.253697 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64ffd\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-kube-api-access-64ffd\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.253776 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-sa-token\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.255189 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-token\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.742466 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.742843 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.748730 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.754945 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver\") pod \"collector-5v5f2\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.766267 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.831952 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-5v5f2" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.945877 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.946233 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-token\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.946467 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-datadir\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.946612 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-datadir" (OuterVolumeSpecName: "datadir") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.946638 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-tmp\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.947109 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-entrypoint\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.947430 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-trusted-ca\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.947637 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.948253 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.948886 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-sa-token\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.949623 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.949992 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64ffd\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-kube-api-access-64ffd\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.950232 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.950448 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config-openshift-service-cacrt\") pod \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\" (UID: \"e981a4b5-8c6c-4d7e-a342-8972a8c69cee\") " Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.950861 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config" (OuterVolumeSpecName: "config") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.950900 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.952134 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.952493 4848 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.952705 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.952863 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.953060 4848 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.953243 4848 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.952867 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-kube-api-access-64ffd" (OuterVolumeSpecName: "kube-api-access-64ffd") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "kube-api-access-64ffd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.953114 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-token" (OuterVolumeSpecName: "collector-token") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.954043 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-tmp" (OuterVolumeSpecName: "tmp") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.954168 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-sa-token" (OuterVolumeSpecName: "sa-token") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.955213 4848 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-datadir\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:47 crc kubenswrapper[4848]: I1204 14:04:47.956105 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics" (OuterVolumeSpecName: "metrics") pod "e981a4b5-8c6c-4d7e-a342-8972a8c69cee" (UID: "e981a4b5-8c6c-4d7e-a342-8972a8c69cee"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.056725 4848 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-collector-token\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.057694 4848 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-tmp\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.057762 4848 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.057820 4848 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.057873 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64ffd\" (UniqueName: \"kubernetes.io/projected/e981a4b5-8c6c-4d7e-a342-8972a8c69cee-kube-api-access-64ffd\") on node \"crc\" DevicePath \"\"" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.775128 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-5v5f2" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.848763 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-5v5f2"] Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.858139 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-5v5f2"] Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.864217 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-wgpm5"] Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.865464 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.868627 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-7nz9k" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.868815 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.870364 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.870681 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-wgpm5"] Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872128 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872183 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-sa-token\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872219 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-metrics\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872254 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-config-openshift-service-cacrt\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872288 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-entrypoint\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872357 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-datadir\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872385 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-config\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872413 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-collector-syslog-receiver\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872431 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnbjg\" (UniqueName: \"kubernetes.io/projected/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-kube-api-access-cnbjg\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872448 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-tmp\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872472 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-collector-token\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.872570 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-trusted-ca\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.873310 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.884111 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973679 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-tmp\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973744 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-collector-token\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973768 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-trusted-ca\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973819 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-sa-token\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973846 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-metrics\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973885 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-config-openshift-service-cacrt\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973917 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-entrypoint\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973961 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-datadir\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.973977 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-config\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.974003 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-collector-syslog-receiver\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.974020 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnbjg\" (UniqueName: \"kubernetes.io/projected/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-kube-api-access-cnbjg\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.974110 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-datadir\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.975137 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-config\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.975471 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-trusted-ca\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.976382 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-entrypoint\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.976370 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-config-openshift-service-cacrt\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.978610 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-collector-token\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.979397 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-collector-syslog-receiver\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.979878 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-tmp\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:48 crc kubenswrapper[4848]: I1204 14:04:48.998463 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-metrics\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:49 crc kubenswrapper[4848]: I1204 14:04:49.005383 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnbjg\" (UniqueName: \"kubernetes.io/projected/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-kube-api-access-cnbjg\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:49 crc kubenswrapper[4848]: I1204 14:04:49.005808 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/13d80a8f-6a05-4fdb-99bd-6aaf5f563f87-sa-token\") pod \"collector-wgpm5\" (UID: \"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87\") " pod="openshift-logging/collector-wgpm5" Dec 04 14:04:49 crc kubenswrapper[4848]: I1204 14:04:49.208942 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-wgpm5" Dec 04 14:04:49 crc kubenswrapper[4848]: I1204 14:04:49.665595 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-wgpm5"] Dec 04 14:04:49 crc kubenswrapper[4848]: I1204 14:04:49.784531 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-wgpm5" event={"ID":"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87","Type":"ContainerStarted","Data":"a0f2d048555a6c19fbbe908755cf5533fb8f710a10d331d30f7bb3ca04c412f5"} Dec 04 14:04:50 crc kubenswrapper[4848]: I1204 14:04:50.401331 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e981a4b5-8c6c-4d7e-a342-8972a8c69cee" path="/var/lib/kubelet/pods/e981a4b5-8c6c-4d7e-a342-8972a8c69cee/volumes" Dec 04 14:04:56 crc kubenswrapper[4848]: I1204 14:04:56.843557 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-wgpm5" event={"ID":"13d80a8f-6a05-4fdb-99bd-6aaf5f563f87","Type":"ContainerStarted","Data":"d79ca4071a683e47696e12137b2ea5d8e10fd70bf3463708f1280c7a84b46fc4"} Dec 04 14:04:56 crc kubenswrapper[4848]: I1204 14:04:56.890269 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-wgpm5" podStartSLOduration=1.963541372 podStartE2EDuration="8.890222935s" podCreationTimestamp="2025-12-04 14:04:48 +0000 UTC" firstStartedPulling="2025-12-04 14:04:49.673361022 +0000 UTC m=+993.615857560" lastFinishedPulling="2025-12-04 14:04:56.600042595 +0000 UTC m=+1000.542539123" observedRunningTime="2025-12-04 14:04:56.876334306 +0000 UTC m=+1000.818830874" watchObservedRunningTime="2025-12-04 14:04:56.890222935 +0000 UTC m=+1000.832719563" Dec 04 14:05:14 crc kubenswrapper[4848]: I1204 14:05:14.314581 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:05:14 crc kubenswrapper[4848]: I1204 14:05:14.315446 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:05:14 crc kubenswrapper[4848]: I1204 14:05:14.315538 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:05:14 crc kubenswrapper[4848]: I1204 14:05:14.316686 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f4368b11f240a04bf6f6e83cbceaec0f4a992be59424fb78d630b078c54c789d"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:05:14 crc kubenswrapper[4848]: I1204 14:05:14.316846 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://f4368b11f240a04bf6f6e83cbceaec0f4a992be59424fb78d630b078c54c789d" gracePeriod=600 Dec 04 14:05:15 crc kubenswrapper[4848]: I1204 14:05:15.007070 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="f4368b11f240a04bf6f6e83cbceaec0f4a992be59424fb78d630b078c54c789d" exitCode=0 Dec 04 14:05:15 crc kubenswrapper[4848]: I1204 14:05:15.007247 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"f4368b11f240a04bf6f6e83cbceaec0f4a992be59424fb78d630b078c54c789d"} Dec 04 14:05:15 crc kubenswrapper[4848]: I1204 14:05:15.007869 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"131123771f065baa008e65d389e63e335b142b4c147179bfa07ab2fb29d26eae"} Dec 04 14:05:15 crc kubenswrapper[4848]: I1204 14:05:15.007901 4848 scope.go:117] "RemoveContainer" containerID="97feeafde380711ec8873168a8a64ca37c03edc0904564e6ca0465fe48907206" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.165378 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr"] Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.167115 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.169124 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.177719 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr"] Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.283895 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.284020 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtg5q\" (UniqueName: \"kubernetes.io/projected/2741f83b-175a-409b-b36e-5ba53521b9b4-kube-api-access-gtg5q\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.284097 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.385720 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.385827 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtg5q\" (UniqueName: \"kubernetes.io/projected/2741f83b-175a-409b-b36e-5ba53521b9b4-kube-api-access-gtg5q\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.385867 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.386333 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.386384 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.414707 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtg5q\" (UniqueName: \"kubernetes.io/projected/2741f83b-175a-409b-b36e-5ba53521b9b4-kube-api-access-gtg5q\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.483361 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:26 crc kubenswrapper[4848]: I1204 14:05:26.728093 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr"] Dec 04 14:05:27 crc kubenswrapper[4848]: I1204 14:05:27.130509 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" event={"ID":"2741f83b-175a-409b-b36e-5ba53521b9b4","Type":"ContainerStarted","Data":"27a15341ca48ed515c37a8cf29feb2543c6cbd9047646a29ab728238c036e3ff"} Dec 04 14:05:28 crc kubenswrapper[4848]: I1204 14:05:28.141119 4848 generic.go:334] "Generic (PLEG): container finished" podID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerID="2521c5da4e99d99ea8a66c2cfbcfa120de1ce1e1bbfddc6380fc94c5ea5dcffd" exitCode=0 Dec 04 14:05:28 crc kubenswrapper[4848]: I1204 14:05:28.141449 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" event={"ID":"2741f83b-175a-409b-b36e-5ba53521b9b4","Type":"ContainerDied","Data":"2521c5da4e99d99ea8a66c2cfbcfa120de1ce1e1bbfddc6380fc94c5ea5dcffd"} Dec 04 14:05:30 crc kubenswrapper[4848]: I1204 14:05:30.165323 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" event={"ID":"2741f83b-175a-409b-b36e-5ba53521b9b4","Type":"ContainerStarted","Data":"a6eea7fd09c40f575b50b8ce3550c1cab84580554b0aee454a2b55450229fb8c"} Dec 04 14:05:31 crc kubenswrapper[4848]: I1204 14:05:31.176242 4848 generic.go:334] "Generic (PLEG): container finished" podID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerID="a6eea7fd09c40f575b50b8ce3550c1cab84580554b0aee454a2b55450229fb8c" exitCode=0 Dec 04 14:05:31 crc kubenswrapper[4848]: I1204 14:05:31.176296 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" event={"ID":"2741f83b-175a-409b-b36e-5ba53521b9b4","Type":"ContainerDied","Data":"a6eea7fd09c40f575b50b8ce3550c1cab84580554b0aee454a2b55450229fb8c"} Dec 04 14:05:32 crc kubenswrapper[4848]: I1204 14:05:32.186708 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" event={"ID":"2741f83b-175a-409b-b36e-5ba53521b9b4","Type":"ContainerStarted","Data":"e76640d686434d67ff79b6afa705f386f24b736cc181dc4dc7ad3afccf1e6f6f"} Dec 04 14:05:32 crc kubenswrapper[4848]: I1204 14:05:32.211974 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" podStartSLOduration=4.40870303 podStartE2EDuration="6.211936942s" podCreationTimestamp="2025-12-04 14:05:26 +0000 UTC" firstStartedPulling="2025-12-04 14:05:28.143834522 +0000 UTC m=+1032.086331090" lastFinishedPulling="2025-12-04 14:05:29.947068484 +0000 UTC m=+1033.889565002" observedRunningTime="2025-12-04 14:05:32.206801666 +0000 UTC m=+1036.149298194" watchObservedRunningTime="2025-12-04 14:05:32.211936942 +0000 UTC m=+1036.154433470" Dec 04 14:05:33 crc kubenswrapper[4848]: I1204 14:05:33.196102 4848 generic.go:334] "Generic (PLEG): container finished" podID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerID="e76640d686434d67ff79b6afa705f386f24b736cc181dc4dc7ad3afccf1e6f6f" exitCode=0 Dec 04 14:05:33 crc kubenswrapper[4848]: I1204 14:05:33.196147 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" event={"ID":"2741f83b-175a-409b-b36e-5ba53521b9b4","Type":"ContainerDied","Data":"e76640d686434d67ff79b6afa705f386f24b736cc181dc4dc7ad3afccf1e6f6f"} Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.507959 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.626560 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-bundle\") pod \"2741f83b-175a-409b-b36e-5ba53521b9b4\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.626692 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtg5q\" (UniqueName: \"kubernetes.io/projected/2741f83b-175a-409b-b36e-5ba53521b9b4-kube-api-access-gtg5q\") pod \"2741f83b-175a-409b-b36e-5ba53521b9b4\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.626786 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-util\") pod \"2741f83b-175a-409b-b36e-5ba53521b9b4\" (UID: \"2741f83b-175a-409b-b36e-5ba53521b9b4\") " Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.627381 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-bundle" (OuterVolumeSpecName: "bundle") pod "2741f83b-175a-409b-b36e-5ba53521b9b4" (UID: "2741f83b-175a-409b-b36e-5ba53521b9b4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.632655 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2741f83b-175a-409b-b36e-5ba53521b9b4-kube-api-access-gtg5q" (OuterVolumeSpecName: "kube-api-access-gtg5q") pod "2741f83b-175a-409b-b36e-5ba53521b9b4" (UID: "2741f83b-175a-409b-b36e-5ba53521b9b4"). InnerVolumeSpecName "kube-api-access-gtg5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.649836 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-util" (OuterVolumeSpecName: "util") pod "2741f83b-175a-409b-b36e-5ba53521b9b4" (UID: "2741f83b-175a-409b-b36e-5ba53521b9b4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.728598 4848 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.728645 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtg5q\" (UniqueName: \"kubernetes.io/projected/2741f83b-175a-409b-b36e-5ba53521b9b4-kube-api-access-gtg5q\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:34 crc kubenswrapper[4848]: I1204 14:05:34.728664 4848 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2741f83b-175a-409b-b36e-5ba53521b9b4-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:05:35 crc kubenswrapper[4848]: I1204 14:05:35.215569 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" event={"ID":"2741f83b-175a-409b-b36e-5ba53521b9b4","Type":"ContainerDied","Data":"27a15341ca48ed515c37a8cf29feb2543c6cbd9047646a29ab728238c036e3ff"} Dec 04 14:05:35 crc kubenswrapper[4848]: I1204 14:05:35.215896 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27a15341ca48ed515c37a8cf29feb2543c6cbd9047646a29ab728238c036e3ff" Dec 04 14:05:35 crc kubenswrapper[4848]: I1204 14:05:35.215671 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr" Dec 04 14:05:37 crc kubenswrapper[4848]: I1204 14:05:37.995091 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9"] Dec 04 14:05:37 crc kubenswrapper[4848]: E1204 14:05:37.995571 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerName="pull" Dec 04 14:05:37 crc kubenswrapper[4848]: I1204 14:05:37.995585 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerName="pull" Dec 04 14:05:37 crc kubenswrapper[4848]: E1204 14:05:37.995613 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerName="util" Dec 04 14:05:37 crc kubenswrapper[4848]: I1204 14:05:37.995622 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerName="util" Dec 04 14:05:37 crc kubenswrapper[4848]: E1204 14:05:37.995641 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerName="extract" Dec 04 14:05:37 crc kubenswrapper[4848]: I1204 14:05:37.995647 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerName="extract" Dec 04 14:05:37 crc kubenswrapper[4848]: I1204 14:05:37.995768 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="2741f83b-175a-409b-b36e-5ba53521b9b4" containerName="extract" Dec 04 14:05:37 crc kubenswrapper[4848]: I1204 14:05:37.996335 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9" Dec 04 14:05:37 crc kubenswrapper[4848]: I1204 14:05:37.999206 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 04 14:05:37 crc kubenswrapper[4848]: I1204 14:05:37.999480 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 04 14:05:38 crc kubenswrapper[4848]: I1204 14:05:38.003699 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-dzwj7" Dec 04 14:05:38 crc kubenswrapper[4848]: I1204 14:05:38.009920 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9"] Dec 04 14:05:38 crc kubenswrapper[4848]: I1204 14:05:38.086043 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f88gf\" (UniqueName: \"kubernetes.io/projected/33caa45e-2061-41cc-8dca-94a382ffa50a-kube-api-access-f88gf\") pod \"nmstate-operator-5b5b58f5c8-69mx9\" (UID: \"33caa45e-2061-41cc-8dca-94a382ffa50a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9" Dec 04 14:05:38 crc kubenswrapper[4848]: I1204 14:05:38.187505 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f88gf\" (UniqueName: \"kubernetes.io/projected/33caa45e-2061-41cc-8dca-94a382ffa50a-kube-api-access-f88gf\") pod \"nmstate-operator-5b5b58f5c8-69mx9\" (UID: \"33caa45e-2061-41cc-8dca-94a382ffa50a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9" Dec 04 14:05:38 crc kubenswrapper[4848]: I1204 14:05:38.208710 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f88gf\" (UniqueName: \"kubernetes.io/projected/33caa45e-2061-41cc-8dca-94a382ffa50a-kube-api-access-f88gf\") pod \"nmstate-operator-5b5b58f5c8-69mx9\" (UID: \"33caa45e-2061-41cc-8dca-94a382ffa50a\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9" Dec 04 14:05:38 crc kubenswrapper[4848]: I1204 14:05:38.311182 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9" Dec 04 14:05:38 crc kubenswrapper[4848]: I1204 14:05:38.801304 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9"] Dec 04 14:05:39 crc kubenswrapper[4848]: I1204 14:05:39.248296 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9" event={"ID":"33caa45e-2061-41cc-8dca-94a382ffa50a","Type":"ContainerStarted","Data":"3978e1037339040e9797699a868aeaee7e2c6e706b2ea5798bfb6ac2aa0aced4"} Dec 04 14:05:42 crc kubenswrapper[4848]: I1204 14:05:42.268612 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9" event={"ID":"33caa45e-2061-41cc-8dca-94a382ffa50a","Type":"ContainerStarted","Data":"379ac9d0b632ce09fc10c7624efcba03ec7bd6c6899505331cc1186ec697b264"} Dec 04 14:05:42 crc kubenswrapper[4848]: I1204 14:05:42.288842 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-69mx9" podStartSLOduration=2.922239626 podStartE2EDuration="5.288818672s" podCreationTimestamp="2025-12-04 14:05:37 +0000 UTC" firstStartedPulling="2025-12-04 14:05:38.797534705 +0000 UTC m=+1042.740031233" lastFinishedPulling="2025-12-04 14:05:41.164113741 +0000 UTC m=+1045.106610279" observedRunningTime="2025-12-04 14:05:42.28423724 +0000 UTC m=+1046.226733768" watchObservedRunningTime="2025-12-04 14:05:42.288818672 +0000 UTC m=+1046.231315210" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.327542 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.329854 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.334510 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.335470 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.335721 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-x9mrv" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.337704 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.343894 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.353277 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.398675 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-rggz7"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.400204 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.435118 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr2zf\" (UniqueName: \"kubernetes.io/projected/ba09ba52-535f-41bc-adb2-210ca566a77a-kube-api-access-qr2zf\") pod \"nmstate-webhook-5f6d4c5ccb-sjwbx\" (UID: \"ba09ba52-535f-41bc-adb2-210ca566a77a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.435205 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ba09ba52-535f-41bc-adb2-210ca566a77a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-sjwbx\" (UID: \"ba09ba52-535f-41bc-adb2-210ca566a77a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.435285 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knqsb\" (UniqueName: \"kubernetes.io/projected/afaef749-1755-4749-bb8f-64b3fcb3f768-kube-api-access-knqsb\") pod \"nmstate-metrics-7f946cbc9-4nd5h\" (UID: \"afaef749-1755-4749-bb8f-64b3fcb3f768\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.504285 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.505354 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.514416 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-8lrzs" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.514680 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.522444 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.526845 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.536753 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knqsb\" (UniqueName: \"kubernetes.io/projected/afaef749-1755-4749-bb8f-64b3fcb3f768-kube-api-access-knqsb\") pod \"nmstate-metrics-7f946cbc9-4nd5h\" (UID: \"afaef749-1755-4749-bb8f-64b3fcb3f768\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.536815 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-ovs-socket\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.536852 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/357ea823-946e-407d-8940-b430c72e324e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.536876 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr2zf\" (UniqueName: \"kubernetes.io/projected/ba09ba52-535f-41bc-adb2-210ca566a77a-kube-api-access-qr2zf\") pod \"nmstate-webhook-5f6d4c5ccb-sjwbx\" (UID: \"ba09ba52-535f-41bc-adb2-210ca566a77a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.536914 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxkt9\" (UniqueName: \"kubernetes.io/projected/357ea823-946e-407d-8940-b430c72e324e-kube-api-access-pxkt9\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.536931 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-nmstate-lock\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.536963 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ba09ba52-535f-41bc-adb2-210ca566a77a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-sjwbx\" (UID: \"ba09ba52-535f-41bc-adb2-210ca566a77a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.536992 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-dbus-socket\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.537013 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/357ea823-946e-407d-8940-b430c72e324e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.537029 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8tz6\" (UniqueName: \"kubernetes.io/projected/973c2430-1dce-45e2-b08f-f87439b80cce-kube-api-access-j8tz6\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: E1204 14:05:47.537148 4848 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 04 14:05:47 crc kubenswrapper[4848]: E1204 14:05:47.537186 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba09ba52-535f-41bc-adb2-210ca566a77a-tls-key-pair podName:ba09ba52-535f-41bc-adb2-210ca566a77a nodeName:}" failed. No retries permitted until 2025-12-04 14:05:48.037170906 +0000 UTC m=+1051.979667434 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/ba09ba52-535f-41bc-adb2-210ca566a77a-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-sjwbx" (UID: "ba09ba52-535f-41bc-adb2-210ca566a77a") : secret "openshift-nmstate-webhook" not found Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.565045 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr2zf\" (UniqueName: \"kubernetes.io/projected/ba09ba52-535f-41bc-adb2-210ca566a77a-kube-api-access-qr2zf\") pod \"nmstate-webhook-5f6d4c5ccb-sjwbx\" (UID: \"ba09ba52-535f-41bc-adb2-210ca566a77a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.572740 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knqsb\" (UniqueName: \"kubernetes.io/projected/afaef749-1755-4749-bb8f-64b3fcb3f768-kube-api-access-knqsb\") pod \"nmstate-metrics-7f946cbc9-4nd5h\" (UID: \"afaef749-1755-4749-bb8f-64b3fcb3f768\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638244 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxkt9\" (UniqueName: \"kubernetes.io/projected/357ea823-946e-407d-8940-b430c72e324e-kube-api-access-pxkt9\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638290 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-nmstate-lock\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638335 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-dbus-socket\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638359 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/357ea823-946e-407d-8940-b430c72e324e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638378 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8tz6\" (UniqueName: \"kubernetes.io/projected/973c2430-1dce-45e2-b08f-f87439b80cce-kube-api-access-j8tz6\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638428 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-ovs-socket\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638460 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/357ea823-946e-407d-8940-b430c72e324e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: E1204 14:05:47.638618 4848 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 04 14:05:47 crc kubenswrapper[4848]: E1204 14:05:47.638670 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/357ea823-946e-407d-8940-b430c72e324e-plugin-serving-cert podName:357ea823-946e-407d-8940-b430c72e324e nodeName:}" failed. No retries permitted until 2025-12-04 14:05:48.138655809 +0000 UTC m=+1052.081152337 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/357ea823-946e-407d-8940-b430c72e324e-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-qj7vf" (UID: "357ea823-946e-407d-8940-b430c72e324e") : secret "plugin-serving-cert" not found Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638923 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-nmstate-lock\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.638923 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-ovs-socket\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.639197 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/973c2430-1dce-45e2-b08f-f87439b80cce-dbus-socket\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.639578 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/357ea823-946e-407d-8940-b430c72e324e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.666385 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.673733 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxkt9\" (UniqueName: \"kubernetes.io/projected/357ea823-946e-407d-8940-b430c72e324e-kube-api-access-pxkt9\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.684152 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8tz6\" (UniqueName: \"kubernetes.io/projected/973c2430-1dce-45e2-b08f-f87439b80cce-kube-api-access-j8tz6\") pod \"nmstate-handler-rggz7\" (UID: \"973c2430-1dce-45e2-b08f-f87439b80cce\") " pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.730744 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.765411 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-8554695464-z4rxq"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.766497 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.776430 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8554695464-z4rxq"] Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.833804 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.841353 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-oauth-config\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.841407 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-service-ca\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.841431 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-trusted-ca-bundle\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.841470 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rds62\" (UniqueName: \"kubernetes.io/projected/457b813b-6604-460c-b16a-64ea6bb703fc-kube-api-access-rds62\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.841487 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-serving-cert\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.841547 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-oauth-serving-cert\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.841571 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-console-config\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.945738 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-service-ca\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.946095 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-trusted-ca-bundle\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.946139 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rds62\" (UniqueName: \"kubernetes.io/projected/457b813b-6604-460c-b16a-64ea6bb703fc-kube-api-access-rds62\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.946155 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-serving-cert\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.946200 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-oauth-serving-cert\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.946219 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-console-config\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.946283 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-oauth-config\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.946670 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-service-ca\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.947517 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-oauth-serving-cert\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.947546 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-trusted-ca-bundle\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.948159 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-console-config\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.951578 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-serving-cert\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.952713 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-oauth-config\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:47 crc kubenswrapper[4848]: I1204 14:05:47.970568 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rds62\" (UniqueName: \"kubernetes.io/projected/457b813b-6604-460c-b16a-64ea6bb703fc-kube-api-access-rds62\") pod \"console-8554695464-z4rxq\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.048802 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ba09ba52-535f-41bc-adb2-210ca566a77a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-sjwbx\" (UID: \"ba09ba52-535f-41bc-adb2-210ca566a77a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.059367 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ba09ba52-535f-41bc-adb2-210ca566a77a-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-sjwbx\" (UID: \"ba09ba52-535f-41bc-adb2-210ca566a77a\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.134405 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.153913 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/357ea823-946e-407d-8940-b430c72e324e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.157609 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/357ea823-946e-407d-8940-b430c72e324e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qj7vf\" (UID: \"357ea823-946e-407d-8940-b430c72e324e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.180985 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h"] Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.282271 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.317461 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" event={"ID":"afaef749-1755-4749-bb8f-64b3fcb3f768","Type":"ContainerStarted","Data":"a5c7b63cdeff8c9270a93a60cc355a272621a32194768485ea5468155de964d8"} Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.318800 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rggz7" event={"ID":"973c2430-1dce-45e2-b08f-f87439b80cce","Type":"ContainerStarted","Data":"29327b27ae0b50a001b87d2c117f55d060166eb5e2874110f19195a6c8965b10"} Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.431845 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.585024 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8554695464-z4rxq"] Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.690113 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx"] Dec 04 14:05:48 crc kubenswrapper[4848]: I1204 14:05:48.880522 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf"] Dec 04 14:05:48 crc kubenswrapper[4848]: W1204 14:05:48.883456 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod357ea823_946e_407d_8940_b430c72e324e.slice/crio-2258e33bad747e4dba855d20337c638118360eb6c50797b56bae2651f5e52a96 WatchSource:0}: Error finding container 2258e33bad747e4dba855d20337c638118360eb6c50797b56bae2651f5e52a96: Status 404 returned error can't find the container with id 2258e33bad747e4dba855d20337c638118360eb6c50797b56bae2651f5e52a96 Dec 04 14:05:49 crc kubenswrapper[4848]: I1204 14:05:49.327291 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" event={"ID":"357ea823-946e-407d-8940-b430c72e324e","Type":"ContainerStarted","Data":"2258e33bad747e4dba855d20337c638118360eb6c50797b56bae2651f5e52a96"} Dec 04 14:05:49 crc kubenswrapper[4848]: I1204 14:05:49.328679 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" event={"ID":"ba09ba52-535f-41bc-adb2-210ca566a77a","Type":"ContainerStarted","Data":"f3b660a54350c7bd73e43aa3e31c502bf907c8ea33bcda2404e5a994520e1e9f"} Dec 04 14:05:49 crc kubenswrapper[4848]: I1204 14:05:49.330446 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8554695464-z4rxq" event={"ID":"457b813b-6604-460c-b16a-64ea6bb703fc","Type":"ContainerStarted","Data":"febbc6b9ab66de76daf5f91b11e58fc53fe3646931b34056207206bb283c64b0"} Dec 04 14:05:49 crc kubenswrapper[4848]: I1204 14:05:49.330504 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8554695464-z4rxq" event={"ID":"457b813b-6604-460c-b16a-64ea6bb703fc","Type":"ContainerStarted","Data":"32611b84f5c8115d7d9cd15f76b2e89d102752cc06249df92c85dfc4197b367d"} Dec 04 14:05:49 crc kubenswrapper[4848]: I1204 14:05:49.353384 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-8554695464-z4rxq" podStartSLOduration=2.353366024 podStartE2EDuration="2.353366024s" podCreationTimestamp="2025-12-04 14:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:05:49.348801883 +0000 UTC m=+1053.291298411" watchObservedRunningTime="2025-12-04 14:05:49.353366024 +0000 UTC m=+1053.295862552" Dec 04 14:05:51 crc kubenswrapper[4848]: I1204 14:05:51.354163 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" event={"ID":"ba09ba52-535f-41bc-adb2-210ca566a77a","Type":"ContainerStarted","Data":"c39cb69e12da2d041816e8dbd90d3ac40ef16be7d782214dfd4c02d60b163531"} Dec 04 14:05:51 crc kubenswrapper[4848]: I1204 14:05:51.355228 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:05:51 crc kubenswrapper[4848]: I1204 14:05:51.356827 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rggz7" event={"ID":"973c2430-1dce-45e2-b08f-f87439b80cce","Type":"ContainerStarted","Data":"bda8e876d411f7836b6300ea6eb43d81bb987488c47c68a5c774568ba4f42ae0"} Dec 04 14:05:51 crc kubenswrapper[4848]: I1204 14:05:51.357014 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:51 crc kubenswrapper[4848]: I1204 14:05:51.359621 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" event={"ID":"afaef749-1755-4749-bb8f-64b3fcb3f768","Type":"ContainerStarted","Data":"0d8d726554c3ed7ffbd86b283b2b626b7d7cf77a7ae69487b74eea00431d74cd"} Dec 04 14:05:51 crc kubenswrapper[4848]: I1204 14:05:51.408858 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-rggz7" podStartSLOduration=1.6205144599999999 podStartE2EDuration="4.40881668s" podCreationTimestamp="2025-12-04 14:05:47 +0000 UTC" firstStartedPulling="2025-12-04 14:05:47.833572358 +0000 UTC m=+1051.776068876" lastFinishedPulling="2025-12-04 14:05:50.621874568 +0000 UTC m=+1054.564371096" observedRunningTime="2025-12-04 14:05:51.391393786 +0000 UTC m=+1055.333890324" watchObservedRunningTime="2025-12-04 14:05:51.40881668 +0000 UTC m=+1055.351313208" Dec 04 14:05:51 crc kubenswrapper[4848]: I1204 14:05:51.413542 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" podStartSLOduration=2.4628264 podStartE2EDuration="4.413525245s" podCreationTimestamp="2025-12-04 14:05:47 +0000 UTC" firstStartedPulling="2025-12-04 14:05:48.698435068 +0000 UTC m=+1052.640931596" lastFinishedPulling="2025-12-04 14:05:50.649133913 +0000 UTC m=+1054.591630441" observedRunningTime="2025-12-04 14:05:51.374692039 +0000 UTC m=+1055.317188567" watchObservedRunningTime="2025-12-04 14:05:51.413525245 +0000 UTC m=+1055.356021773" Dec 04 14:05:52 crc kubenswrapper[4848]: I1204 14:05:52.372536 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" event={"ID":"357ea823-946e-407d-8940-b430c72e324e","Type":"ContainerStarted","Data":"a7c631d47f508e3765253a3d4d250d7c4e3af6407bff77227d6a17030ea46975"} Dec 04 14:05:52 crc kubenswrapper[4848]: I1204 14:05:52.399732 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qj7vf" podStartSLOduration=2.643416111 podStartE2EDuration="5.399700341s" podCreationTimestamp="2025-12-04 14:05:47 +0000 UTC" firstStartedPulling="2025-12-04 14:05:48.88609273 +0000 UTC m=+1052.828589258" lastFinishedPulling="2025-12-04 14:05:51.64237696 +0000 UTC m=+1055.584873488" observedRunningTime="2025-12-04 14:05:52.390702782 +0000 UTC m=+1056.333199330" watchObservedRunningTime="2025-12-04 14:05:52.399700341 +0000 UTC m=+1056.342196869" Dec 04 14:05:54 crc kubenswrapper[4848]: I1204 14:05:54.392181 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" event={"ID":"afaef749-1755-4749-bb8f-64b3fcb3f768","Type":"ContainerStarted","Data":"568f6929412907fa50b8093539fd06777a70961128fbdf9da451ebe900e4cd42"} Dec 04 14:05:54 crc kubenswrapper[4848]: I1204 14:05:54.418064 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-4nd5h" podStartSLOduration=2.418362608 podStartE2EDuration="7.418039724s" podCreationTimestamp="2025-12-04 14:05:47 +0000 UTC" firstStartedPulling="2025-12-04 14:05:48.204811872 +0000 UTC m=+1052.147308400" lastFinishedPulling="2025-12-04 14:05:53.204488988 +0000 UTC m=+1057.146985516" observedRunningTime="2025-12-04 14:05:54.410427808 +0000 UTC m=+1058.352924356" watchObservedRunningTime="2025-12-04 14:05:54.418039724 +0000 UTC m=+1058.360536292" Dec 04 14:05:57 crc kubenswrapper[4848]: I1204 14:05:57.764849 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-rggz7" Dec 04 14:05:58 crc kubenswrapper[4848]: I1204 14:05:58.135198 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:58 crc kubenswrapper[4848]: I1204 14:05:58.135902 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:58 crc kubenswrapper[4848]: I1204 14:05:58.145087 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:58 crc kubenswrapper[4848]: I1204 14:05:58.432945 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:05:58 crc kubenswrapper[4848]: I1204 14:05:58.518424 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-74cc9ddcc4-dx5vs"] Dec 04 14:06:08 crc kubenswrapper[4848]: I1204 14:06:08.288654 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-sjwbx" Dec 04 14:06:23 crc kubenswrapper[4848]: I1204 14:06:23.567829 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-74cc9ddcc4-dx5vs" podUID="9227885e-0ea7-47fa-a763-3a3dd3c1de39" containerName="console" containerID="cri-o://4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094" gracePeriod=15 Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.019630 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74cc9ddcc4-dx5vs_9227885e-0ea7-47fa-a763-3a3dd3c1de39/console/0.log" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.019925 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.117555 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jg2v\" (UniqueName: \"kubernetes.io/projected/9227885e-0ea7-47fa-a763-3a3dd3c1de39-kube-api-access-7jg2v\") pod \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.117621 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-service-ca\") pod \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.117667 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-oauth-config\") pod \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.117716 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-config\") pod \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.117733 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-trusted-ca-bundle\") pod \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.117767 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-oauth-serving-cert\") pod \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.117808 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-serving-cert\") pod \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\" (UID: \"9227885e-0ea7-47fa-a763-3a3dd3c1de39\") " Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.122671 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9227885e-0ea7-47fa-a763-3a3dd3c1de39" (UID: "9227885e-0ea7-47fa-a763-3a3dd3c1de39"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.122983 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-config" (OuterVolumeSpecName: "console-config") pod "9227885e-0ea7-47fa-a763-3a3dd3c1de39" (UID: "9227885e-0ea7-47fa-a763-3a3dd3c1de39"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.123202 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9227885e-0ea7-47fa-a763-3a3dd3c1de39" (UID: "9227885e-0ea7-47fa-a763-3a3dd3c1de39"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.124702 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-service-ca" (OuterVolumeSpecName: "service-ca") pod "9227885e-0ea7-47fa-a763-3a3dd3c1de39" (UID: "9227885e-0ea7-47fa-a763-3a3dd3c1de39"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.134630 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9227885e-0ea7-47fa-a763-3a3dd3c1de39" (UID: "9227885e-0ea7-47fa-a763-3a3dd3c1de39"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.136180 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9227885e-0ea7-47fa-a763-3a3dd3c1de39-kube-api-access-7jg2v" (OuterVolumeSpecName: "kube-api-access-7jg2v") pod "9227885e-0ea7-47fa-a763-3a3dd3c1de39" (UID: "9227885e-0ea7-47fa-a763-3a3dd3c1de39"). InnerVolumeSpecName "kube-api-access-7jg2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.143464 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9227885e-0ea7-47fa-a763-3a3dd3c1de39" (UID: "9227885e-0ea7-47fa-a763-3a3dd3c1de39"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.219668 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jg2v\" (UniqueName: \"kubernetes.io/projected/9227885e-0ea7-47fa-a763-3a3dd3c1de39-kube-api-access-7jg2v\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.219712 4848 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.219731 4848 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.219744 4848 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.219756 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.219767 4848 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9227885e-0ea7-47fa-a763-3a3dd3c1de39-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.219779 4848 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9227885e-0ea7-47fa-a763-3a3dd3c1de39-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.654553 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74cc9ddcc4-dx5vs_9227885e-0ea7-47fa-a763-3a3dd3c1de39/console/0.log" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.654828 4848 generic.go:334] "Generic (PLEG): container finished" podID="9227885e-0ea7-47fa-a763-3a3dd3c1de39" containerID="4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094" exitCode=2 Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.654860 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74cc9ddcc4-dx5vs" event={"ID":"9227885e-0ea7-47fa-a763-3a3dd3c1de39","Type":"ContainerDied","Data":"4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094"} Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.654894 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74cc9ddcc4-dx5vs" event={"ID":"9227885e-0ea7-47fa-a763-3a3dd3c1de39","Type":"ContainerDied","Data":"d5f63630cda22165cfe0c5d5846d9bbe3f7c614af160ec66c8d41bb20e7daadf"} Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.654895 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74cc9ddcc4-dx5vs" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.654927 4848 scope.go:117] "RemoveContainer" containerID="4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.688081 4848 scope.go:117] "RemoveContainer" containerID="4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094" Dec 04 14:06:24 crc kubenswrapper[4848]: E1204 14:06:24.688898 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094\": container with ID starting with 4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094 not found: ID does not exist" containerID="4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.688941 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094"} err="failed to get container status \"4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094\": rpc error: code = NotFound desc = could not find container \"4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094\": container with ID starting with 4f83e14ec179d480946fac4cb305bd4a5884cc3b16ff28ac6ca0d94eacf36094 not found: ID does not exist" Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.689893 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-74cc9ddcc4-dx5vs"] Dec 04 14:06:24 crc kubenswrapper[4848]: I1204 14:06:24.695226 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-74cc9ddcc4-dx5vs"] Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.434433 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn"] Dec 04 14:06:25 crc kubenswrapper[4848]: E1204 14:06:25.434690 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9227885e-0ea7-47fa-a763-3a3dd3c1de39" containerName="console" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.434701 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9227885e-0ea7-47fa-a763-3a3dd3c1de39" containerName="console" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.434835 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9227885e-0ea7-47fa-a763-3a3dd3c1de39" containerName="console" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.435758 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.438169 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.438439 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s5dd\" (UniqueName: \"kubernetes.io/projected/9fa52fcb-11a4-4729-b5b8-4ce27d912310-kube-api-access-5s5dd\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.438496 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.438593 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.446895 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn"] Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.539799 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.540017 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s5dd\" (UniqueName: \"kubernetes.io/projected/9fa52fcb-11a4-4729-b5b8-4ce27d912310-kube-api-access-5s5dd\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.540087 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.540581 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.540605 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.563046 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s5dd\" (UniqueName: \"kubernetes.io/projected/9fa52fcb-11a4-4729-b5b8-4ce27d912310-kube-api-access-5s5dd\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:25 crc kubenswrapper[4848]: I1204 14:06:25.753079 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:26 crc kubenswrapper[4848]: I1204 14:06:26.170838 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn"] Dec 04 14:06:26 crc kubenswrapper[4848]: I1204 14:06:26.407530 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9227885e-0ea7-47fa-a763-3a3dd3c1de39" path="/var/lib/kubelet/pods/9227885e-0ea7-47fa-a763-3a3dd3c1de39/volumes" Dec 04 14:06:26 crc kubenswrapper[4848]: I1204 14:06:26.673292 4848 generic.go:334] "Generic (PLEG): container finished" podID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerID="750b2d92df21c8dcc50154799dbe475c07c2019c3b473823b83617d221de2241" exitCode=0 Dec 04 14:06:26 crc kubenswrapper[4848]: I1204 14:06:26.673340 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" event={"ID":"9fa52fcb-11a4-4729-b5b8-4ce27d912310","Type":"ContainerDied","Data":"750b2d92df21c8dcc50154799dbe475c07c2019c3b473823b83617d221de2241"} Dec 04 14:06:26 crc kubenswrapper[4848]: I1204 14:06:26.673369 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" event={"ID":"9fa52fcb-11a4-4729-b5b8-4ce27d912310","Type":"ContainerStarted","Data":"04ab683f8b3c30235eefeff15b76fd3f8bdee341b082007722789d7f5fc5d7a7"} Dec 04 14:06:28 crc kubenswrapper[4848]: I1204 14:06:28.693927 4848 generic.go:334] "Generic (PLEG): container finished" podID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerID="7515addfc01d0d3874a79d4156bdfafe8f741c2bc4e1178d696c116b1c3006a6" exitCode=0 Dec 04 14:06:28 crc kubenswrapper[4848]: I1204 14:06:28.694077 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" event={"ID":"9fa52fcb-11a4-4729-b5b8-4ce27d912310","Type":"ContainerDied","Data":"7515addfc01d0d3874a79d4156bdfafe8f741c2bc4e1178d696c116b1c3006a6"} Dec 04 14:06:29 crc kubenswrapper[4848]: I1204 14:06:29.704371 4848 generic.go:334] "Generic (PLEG): container finished" podID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerID="a9bd65dcdc5332726ccccd2476434a43dee073a9723b3bec2eb101fc27884be5" exitCode=0 Dec 04 14:06:29 crc kubenswrapper[4848]: I1204 14:06:29.704482 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" event={"ID":"9fa52fcb-11a4-4729-b5b8-4ce27d912310","Type":"ContainerDied","Data":"a9bd65dcdc5332726ccccd2476434a43dee073a9723b3bec2eb101fc27884be5"} Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.003807 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.124192 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s5dd\" (UniqueName: \"kubernetes.io/projected/9fa52fcb-11a4-4729-b5b8-4ce27d912310-kube-api-access-5s5dd\") pod \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.124314 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-bundle\") pod \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.124466 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-util\") pod \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\" (UID: \"9fa52fcb-11a4-4729-b5b8-4ce27d912310\") " Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.125364 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-bundle" (OuterVolumeSpecName: "bundle") pod "9fa52fcb-11a4-4729-b5b8-4ce27d912310" (UID: "9fa52fcb-11a4-4729-b5b8-4ce27d912310"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.131791 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fa52fcb-11a4-4729-b5b8-4ce27d912310-kube-api-access-5s5dd" (OuterVolumeSpecName: "kube-api-access-5s5dd") pod "9fa52fcb-11a4-4729-b5b8-4ce27d912310" (UID: "9fa52fcb-11a4-4729-b5b8-4ce27d912310"). InnerVolumeSpecName "kube-api-access-5s5dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.149694 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-util" (OuterVolumeSpecName: "util") pod "9fa52fcb-11a4-4729-b5b8-4ce27d912310" (UID: "9fa52fcb-11a4-4729-b5b8-4ce27d912310"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.225929 4848 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.225975 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s5dd\" (UniqueName: \"kubernetes.io/projected/9fa52fcb-11a4-4729-b5b8-4ce27d912310-kube-api-access-5s5dd\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.225987 4848 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa52fcb-11a4-4729-b5b8-4ce27d912310-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.719886 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" event={"ID":"9fa52fcb-11a4-4729-b5b8-4ce27d912310","Type":"ContainerDied","Data":"04ab683f8b3c30235eefeff15b76fd3f8bdee341b082007722789d7f5fc5d7a7"} Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.719933 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04ab683f8b3c30235eefeff15b76fd3f8bdee341b082007722789d7f5fc5d7a7" Dec 04 14:06:31 crc kubenswrapper[4848]: I1204 14:06:31.719971 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.358014 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns"] Dec 04 14:06:43 crc kubenswrapper[4848]: E1204 14:06:43.360053 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerName="util" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.360153 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerName="util" Dec 04 14:06:43 crc kubenswrapper[4848]: E1204 14:06:43.360247 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerName="extract" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.360314 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerName="extract" Dec 04 14:06:43 crc kubenswrapper[4848]: E1204 14:06:43.360386 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerName="pull" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.360459 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerName="pull" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.360719 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fa52fcb-11a4-4729-b5b8-4ce27d912310" containerName="extract" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.361456 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.365259 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.366067 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.366102 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.366271 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-hlh6c" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.369278 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.378175 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns"] Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.519179 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh4pg\" (UniqueName: \"kubernetes.io/projected/fe476aa8-681e-4dfb-9e65-58910c24c9cc-kube-api-access-dh4pg\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.520124 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe476aa8-681e-4dfb-9e65-58910c24c9cc-apiservice-cert\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.520551 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe476aa8-681e-4dfb-9e65-58910c24c9cc-webhook-cert\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.621342 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe476aa8-681e-4dfb-9e65-58910c24c9cc-webhook-cert\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.621425 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh4pg\" (UniqueName: \"kubernetes.io/projected/fe476aa8-681e-4dfb-9e65-58910c24c9cc-kube-api-access-dh4pg\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.621489 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe476aa8-681e-4dfb-9e65-58910c24c9cc-apiservice-cert\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.627670 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe476aa8-681e-4dfb-9e65-58910c24c9cc-webhook-cert\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.627671 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe476aa8-681e-4dfb-9e65-58910c24c9cc-apiservice-cert\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.638666 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh4pg\" (UniqueName: \"kubernetes.io/projected/fe476aa8-681e-4dfb-9e65-58910c24c9cc-kube-api-access-dh4pg\") pod \"metallb-operator-controller-manager-657d4c7649-z2tns\" (UID: \"fe476aa8-681e-4dfb-9e65-58910c24c9cc\") " pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.681461 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.705591 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg"] Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.706668 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.722166 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.722312 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-j8r6g" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.722377 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.754778 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg"] Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.825484 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab8ac511-522e-4116-9c7b-ab61a4a6e018-apiservice-cert\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.825908 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab8ac511-522e-4116-9c7b-ab61a4a6e018-webhook-cert\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.825975 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbnh5\" (UniqueName: \"kubernetes.io/projected/ab8ac511-522e-4116-9c7b-ab61a4a6e018-kube-api-access-rbnh5\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.929049 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab8ac511-522e-4116-9c7b-ab61a4a6e018-apiservice-cert\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.929153 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab8ac511-522e-4116-9c7b-ab61a4a6e018-webhook-cert\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.929195 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbnh5\" (UniqueName: \"kubernetes.io/projected/ab8ac511-522e-4116-9c7b-ab61a4a6e018-kube-api-access-rbnh5\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.953929 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ab8ac511-522e-4116-9c7b-ab61a4a6e018-webhook-cert\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.955755 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ab8ac511-522e-4116-9c7b-ab61a4a6e018-apiservice-cert\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:43 crc kubenswrapper[4848]: I1204 14:06:43.957276 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbnh5\" (UniqueName: \"kubernetes.io/projected/ab8ac511-522e-4116-9c7b-ab61a4a6e018-kube-api-access-rbnh5\") pod \"metallb-operator-webhook-server-c56865d68-9mdfg\" (UID: \"ab8ac511-522e-4116-9c7b-ab61a4a6e018\") " pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:44 crc kubenswrapper[4848]: I1204 14:06:44.071330 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:06:44 crc kubenswrapper[4848]: W1204 14:06:44.272288 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe476aa8_681e_4dfb_9e65_58910c24c9cc.slice/crio-71a1ad8904dea21d1e349e6ebefaf3ccfe23cc69b8a44f1c60a3da27dd70204d WatchSource:0}: Error finding container 71a1ad8904dea21d1e349e6ebefaf3ccfe23cc69b8a44f1c60a3da27dd70204d: Status 404 returned error can't find the container with id 71a1ad8904dea21d1e349e6ebefaf3ccfe23cc69b8a44f1c60a3da27dd70204d Dec 04 14:06:44 crc kubenswrapper[4848]: I1204 14:06:44.281168 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns"] Dec 04 14:06:44 crc kubenswrapper[4848]: I1204 14:06:44.670122 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg"] Dec 04 14:06:44 crc kubenswrapper[4848]: I1204 14:06:44.853516 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" event={"ID":"fe476aa8-681e-4dfb-9e65-58910c24c9cc","Type":"ContainerStarted","Data":"71a1ad8904dea21d1e349e6ebefaf3ccfe23cc69b8a44f1c60a3da27dd70204d"} Dec 04 14:06:44 crc kubenswrapper[4848]: I1204 14:06:44.855061 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" event={"ID":"ab8ac511-522e-4116-9c7b-ab61a4a6e018","Type":"ContainerStarted","Data":"1b263770c9b44f1c64b484a793ee0494b9306a423f30c702dd7c16326e1d54d1"} Dec 04 14:06:47 crc kubenswrapper[4848]: I1204 14:06:47.880686 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" event={"ID":"fe476aa8-681e-4dfb-9e65-58910c24c9cc","Type":"ContainerStarted","Data":"33508a8a597a1e1ad7e2c25674cb88bcd4bf8cb009ebfd8d23404e5d3564504a"} Dec 04 14:06:47 crc kubenswrapper[4848]: I1204 14:06:47.882403 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:06:49 crc kubenswrapper[4848]: I1204 14:06:49.898618 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" event={"ID":"ab8ac511-522e-4116-9c7b-ab61a4a6e018","Type":"ContainerStarted","Data":"b4dd95d461b0ce809de8da220733f001073562538ef2d6827c86794011693df6"} Dec 04 14:06:49 crc kubenswrapper[4848]: I1204 14:06:49.928926 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" podStartSLOduration=2.193003276 podStartE2EDuration="6.928908336s" podCreationTimestamp="2025-12-04 14:06:43 +0000 UTC" firstStartedPulling="2025-12-04 14:06:44.672700756 +0000 UTC m=+1108.615197284" lastFinishedPulling="2025-12-04 14:06:49.408605826 +0000 UTC m=+1113.351102344" observedRunningTime="2025-12-04 14:06:49.925275199 +0000 UTC m=+1113.867771767" watchObservedRunningTime="2025-12-04 14:06:49.928908336 +0000 UTC m=+1113.871404874" Dec 04 14:06:49 crc kubenswrapper[4848]: I1204 14:06:49.930451 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" podStartSLOduration=3.776405371 podStartE2EDuration="6.930437503s" podCreationTimestamp="2025-12-04 14:06:43 +0000 UTC" firstStartedPulling="2025-12-04 14:06:44.280506805 +0000 UTC m=+1108.223003333" lastFinishedPulling="2025-12-04 14:06:47.434538937 +0000 UTC m=+1111.377035465" observedRunningTime="2025-12-04 14:06:47.907772219 +0000 UTC m=+1111.850268737" watchObservedRunningTime="2025-12-04 14:06:49.930437503 +0000 UTC m=+1113.872934051" Dec 04 14:06:50 crc kubenswrapper[4848]: I1204 14:06:50.907102 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:07:04 crc kubenswrapper[4848]: I1204 14:07:04.083136 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-c56865d68-9mdfg" Dec 04 14:07:14 crc kubenswrapper[4848]: I1204 14:07:14.313795 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:07:14 crc kubenswrapper[4848]: I1204 14:07:14.314400 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:07:23 crc kubenswrapper[4848]: I1204 14:07:23.685228 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-657d4c7649-z2tns" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.433362 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-j45lv"] Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.438510 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.452916 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr"] Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.453878 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.459642 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ddv5f" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.459809 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.459906 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.460025 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.531357 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr"] Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539350 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsn9b\" (UniqueName: \"kubernetes.io/projected/054c1d24-8144-490d-b917-391860e5704d-kube-api-access-rsn9b\") pod \"frr-k8s-webhook-server-7fcb986d4-gpdjr\" (UID: \"054c1d24-8144-490d-b917-391860e5704d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539387 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054c1d24-8144-490d-b917-391860e5704d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-gpdjr\" (UID: \"054c1d24-8144-490d-b917-391860e5704d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539444 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-conf\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539483 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-startup\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539500 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-reloader\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539525 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-sockets\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539539 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrt7c\" (UniqueName: \"kubernetes.io/projected/a5455d26-d1da-4e85-a556-a14cba35f8a9-kube-api-access-rrt7c\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539580 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-metrics\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.539600 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5455d26-d1da-4e85-a556-a14cba35f8a9-metrics-certs\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.564685 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-pxcmt"] Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.566644 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.568831 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.569225 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-q62lf" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.569242 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.569290 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.587001 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-h8bgt"] Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.588577 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.590018 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.615043 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-h8bgt"] Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642667 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-metrics\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642712 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5455d26-d1da-4e85-a556-a14cba35f8a9-metrics-certs\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642763 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642788 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsn9b\" (UniqueName: \"kubernetes.io/projected/054c1d24-8144-490d-b917-391860e5704d-kube-api-access-rsn9b\") pod \"frr-k8s-webhook-server-7fcb986d4-gpdjr\" (UID: \"054c1d24-8144-490d-b917-391860e5704d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642807 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054c1d24-8144-490d-b917-391860e5704d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-gpdjr\" (UID: \"054c1d24-8144-490d-b917-391860e5704d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642836 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzl49\" (UniqueName: \"kubernetes.io/projected/e1998e0d-8a5c-4a7a-8008-598967ed2da0-kube-api-access-zzl49\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642858 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-conf\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642882 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-startup\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642897 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-reloader\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642914 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metallb-excludel2\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642931 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-sockets\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642958 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrt7c\" (UniqueName: \"kubernetes.io/projected/a5455d26-d1da-4e85-a556-a14cba35f8a9-kube-api-access-rrt7c\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.642987 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metrics-certs\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.643361 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-metrics\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.644443 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-conf\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.644757 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-reloader\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.644979 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-sockets\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.645481 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a5455d26-d1da-4e85-a556-a14cba35f8a9-frr-startup\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.663827 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054c1d24-8144-490d-b917-391860e5704d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-gpdjr\" (UID: \"054c1d24-8144-490d-b917-391860e5704d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.665037 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5455d26-d1da-4e85-a556-a14cba35f8a9-metrics-certs\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.667639 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsn9b\" (UniqueName: \"kubernetes.io/projected/054c1d24-8144-490d-b917-391860e5704d-kube-api-access-rsn9b\") pod \"frr-k8s-webhook-server-7fcb986d4-gpdjr\" (UID: \"054c1d24-8144-490d-b917-391860e5704d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.669215 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrt7c\" (UniqueName: \"kubernetes.io/projected/a5455d26-d1da-4e85-a556-a14cba35f8a9-kube-api-access-rrt7c\") pod \"frr-k8s-j45lv\" (UID: \"a5455d26-d1da-4e85-a556-a14cba35f8a9\") " pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.744323 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metallb-excludel2\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.744394 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metrics-certs\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.744470 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.744493 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c600d3-9a0f-4cab-9581-e75fa003c19d-metrics-certs\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.744526 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8c600d3-9a0f-4cab-9581-e75fa003c19d-cert\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.744562 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzl49\" (UniqueName: \"kubernetes.io/projected/e1998e0d-8a5c-4a7a-8008-598967ed2da0-kube-api-access-zzl49\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.744603 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpfdt\" (UniqueName: \"kubernetes.io/projected/e8c600d3-9a0f-4cab-9581-e75fa003c19d-kube-api-access-qpfdt\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: E1204 14:07:24.744616 4848 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 04 14:07:24 crc kubenswrapper[4848]: E1204 14:07:24.744687 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metrics-certs podName:e1998e0d-8a5c-4a7a-8008-598967ed2da0 nodeName:}" failed. No retries permitted until 2025-12-04 14:07:25.244669664 +0000 UTC m=+1149.187166192 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metrics-certs") pod "speaker-pxcmt" (UID: "e1998e0d-8a5c-4a7a-8008-598967ed2da0") : secret "speaker-certs-secret" not found Dec 04 14:07:24 crc kubenswrapper[4848]: E1204 14:07:24.744786 4848 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 14:07:24 crc kubenswrapper[4848]: E1204 14:07:24.744851 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist podName:e1998e0d-8a5c-4a7a-8008-598967ed2da0 nodeName:}" failed. No retries permitted until 2025-12-04 14:07:25.244835157 +0000 UTC m=+1149.187331685 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist") pod "speaker-pxcmt" (UID: "e1998e0d-8a5c-4a7a-8008-598967ed2da0") : secret "metallb-memberlist" not found Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.745183 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metallb-excludel2\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.758568 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.763473 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzl49\" (UniqueName: \"kubernetes.io/projected/e1998e0d-8a5c-4a7a-8008-598967ed2da0-kube-api-access-zzl49\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.774838 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.846071 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c600d3-9a0f-4cab-9581-e75fa003c19d-metrics-certs\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.846132 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8c600d3-9a0f-4cab-9581-e75fa003c19d-cert\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.846201 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpfdt\" (UniqueName: \"kubernetes.io/projected/e8c600d3-9a0f-4cab-9581-e75fa003c19d-kube-api-access-qpfdt\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.847984 4848 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.851078 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e8c600d3-9a0f-4cab-9581-e75fa003c19d-metrics-certs\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.861938 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e8c600d3-9a0f-4cab-9581-e75fa003c19d-cert\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.868024 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpfdt\" (UniqueName: \"kubernetes.io/projected/e8c600d3-9a0f-4cab-9581-e75fa003c19d-kube-api-access-qpfdt\") pod \"controller-f8648f98b-h8bgt\" (UID: \"e8c600d3-9a0f-4cab-9581-e75fa003c19d\") " pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:24 crc kubenswrapper[4848]: I1204 14:07:24.908067 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:25 crc kubenswrapper[4848]: I1204 14:07:25.197720 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr"] Dec 04 14:07:25 crc kubenswrapper[4848]: I1204 14:07:25.202144 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerStarted","Data":"5a76277e5dadab6b1f10cc0802ac556a990ad6933e3a0ed5257a0ebe1cfb5f1d"} Dec 04 14:07:25 crc kubenswrapper[4848]: W1204 14:07:25.205533 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod054c1d24_8144_490d_b917_391860e5704d.slice/crio-598fd31fa4cc9bea31a36557750fd1f9b0e3518e2eb0fa3f083ec7f8746c46cb WatchSource:0}: Error finding container 598fd31fa4cc9bea31a36557750fd1f9b0e3518e2eb0fa3f083ec7f8746c46cb: Status 404 returned error can't find the container with id 598fd31fa4cc9bea31a36557750fd1f9b0e3518e2eb0fa3f083ec7f8746c46cb Dec 04 14:07:25 crc kubenswrapper[4848]: I1204 14:07:25.252253 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metrics-certs\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:25 crc kubenswrapper[4848]: I1204 14:07:25.252696 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:25 crc kubenswrapper[4848]: E1204 14:07:25.252975 4848 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 14:07:25 crc kubenswrapper[4848]: E1204 14:07:25.253093 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist podName:e1998e0d-8a5c-4a7a-8008-598967ed2da0 nodeName:}" failed. No retries permitted until 2025-12-04 14:07:26.253070476 +0000 UTC m=+1150.195567004 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist") pod "speaker-pxcmt" (UID: "e1998e0d-8a5c-4a7a-8008-598967ed2da0") : secret "metallb-memberlist" not found Dec 04 14:07:25 crc kubenswrapper[4848]: I1204 14:07:25.256125 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-metrics-certs\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:25 crc kubenswrapper[4848]: I1204 14:07:25.380829 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-h8bgt"] Dec 04 14:07:25 crc kubenswrapper[4848]: W1204 14:07:25.390413 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8c600d3_9a0f_4cab_9581_e75fa003c19d.slice/crio-2cdd4c3b906cc2372ea8b4037126bcdd7d525d800802dc432c09452f2ba9e385 WatchSource:0}: Error finding container 2cdd4c3b906cc2372ea8b4037126bcdd7d525d800802dc432c09452f2ba9e385: Status 404 returned error can't find the container with id 2cdd4c3b906cc2372ea8b4037126bcdd7d525d800802dc432c09452f2ba9e385 Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.213046 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" event={"ID":"054c1d24-8144-490d-b917-391860e5704d","Type":"ContainerStarted","Data":"598fd31fa4cc9bea31a36557750fd1f9b0e3518e2eb0fa3f083ec7f8746c46cb"} Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.216374 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-h8bgt" event={"ID":"e8c600d3-9a0f-4cab-9581-e75fa003c19d","Type":"ContainerStarted","Data":"b267a612ca84b71bd5e38dd1cb22d557127ac94a2ac2c4774c925fdb246bed30"} Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.216587 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-h8bgt" event={"ID":"e8c600d3-9a0f-4cab-9581-e75fa003c19d","Type":"ContainerStarted","Data":"d961d2011aa361946046bcf8a615e8659ea6759f708ff5a1a6cd6425dae61b2c"} Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.216769 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.216937 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-h8bgt" event={"ID":"e8c600d3-9a0f-4cab-9581-e75fa003c19d","Type":"ContainerStarted","Data":"2cdd4c3b906cc2372ea8b4037126bcdd7d525d800802dc432c09452f2ba9e385"} Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.269046 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.277204 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1998e0d-8a5c-4a7a-8008-598967ed2da0-memberlist\") pod \"speaker-pxcmt\" (UID: \"e1998e0d-8a5c-4a7a-8008-598967ed2da0\") " pod="metallb-system/speaker-pxcmt" Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.387456 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-pxcmt" Dec 04 14:07:26 crc kubenswrapper[4848]: W1204 14:07:26.411018 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1998e0d_8a5c_4a7a_8008_598967ed2da0.slice/crio-539b5bc3b9cbb1561b162af4be03ca13e33d1b82e53f79ee548b17e5605d8d06 WatchSource:0}: Error finding container 539b5bc3b9cbb1561b162af4be03ca13e33d1b82e53f79ee548b17e5605d8d06: Status 404 returned error can't find the container with id 539b5bc3b9cbb1561b162af4be03ca13e33d1b82e53f79ee548b17e5605d8d06 Dec 04 14:07:26 crc kubenswrapper[4848]: I1204 14:07:26.426303 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-h8bgt" podStartSLOduration=2.426281976 podStartE2EDuration="2.426281976s" podCreationTimestamp="2025-12-04 14:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:07:26.241611317 +0000 UTC m=+1150.184107895" watchObservedRunningTime="2025-12-04 14:07:26.426281976 +0000 UTC m=+1150.368778514" Dec 04 14:07:27 crc kubenswrapper[4848]: I1204 14:07:27.229957 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pxcmt" event={"ID":"e1998e0d-8a5c-4a7a-8008-598967ed2da0","Type":"ContainerStarted","Data":"86fd5ff1abd6f221b1a03c87c4d8b4fdfc1267374ae9990d86d98dccf6275dd4"} Dec 04 14:07:27 crc kubenswrapper[4848]: I1204 14:07:27.230290 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pxcmt" event={"ID":"e1998e0d-8a5c-4a7a-8008-598967ed2da0","Type":"ContainerStarted","Data":"6080bea91aa07fb5ead7f96e519adf8b1989cb44e401159a72b130034da65787"} Dec 04 14:07:27 crc kubenswrapper[4848]: I1204 14:07:27.230303 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-pxcmt" event={"ID":"e1998e0d-8a5c-4a7a-8008-598967ed2da0","Type":"ContainerStarted","Data":"539b5bc3b9cbb1561b162af4be03ca13e33d1b82e53f79ee548b17e5605d8d06"} Dec 04 14:07:27 crc kubenswrapper[4848]: I1204 14:07:27.230451 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-pxcmt" Dec 04 14:07:34 crc kubenswrapper[4848]: I1204 14:07:34.297268 4848 generic.go:334] "Generic (PLEG): container finished" podID="a5455d26-d1da-4e85-a556-a14cba35f8a9" containerID="4bdeb319865a2116469115a05ff1e9d84389896851d08c7838ed9efcd9dfee64" exitCode=0 Dec 04 14:07:34 crc kubenswrapper[4848]: I1204 14:07:34.297319 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerDied","Data":"4bdeb319865a2116469115a05ff1e9d84389896851d08c7838ed9efcd9dfee64"} Dec 04 14:07:34 crc kubenswrapper[4848]: I1204 14:07:34.299517 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" event={"ID":"054c1d24-8144-490d-b917-391860e5704d","Type":"ContainerStarted","Data":"3d0e814e439f816ecf5b116eeeb05f31cc421a8cf3dc0edfb0675e49e6be6b49"} Dec 04 14:07:34 crc kubenswrapper[4848]: I1204 14:07:34.299735 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:34 crc kubenswrapper[4848]: I1204 14:07:34.335924 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-pxcmt" podStartSLOduration=10.335904964 podStartE2EDuration="10.335904964s" podCreationTimestamp="2025-12-04 14:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:07:27.251371861 +0000 UTC m=+1151.193868389" watchObservedRunningTime="2025-12-04 14:07:34.335904964 +0000 UTC m=+1158.278401502" Dec 04 14:07:34 crc kubenswrapper[4848]: I1204 14:07:34.348927 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" podStartSLOduration=3.044546247 podStartE2EDuration="10.348901789s" podCreationTimestamp="2025-12-04 14:07:24 +0000 UTC" firstStartedPulling="2025-12-04 14:07:25.207537335 +0000 UTC m=+1149.150033863" lastFinishedPulling="2025-12-04 14:07:32.511892877 +0000 UTC m=+1156.454389405" observedRunningTime="2025-12-04 14:07:34.345102387 +0000 UTC m=+1158.287598955" watchObservedRunningTime="2025-12-04 14:07:34.348901789 +0000 UTC m=+1158.291398317" Dec 04 14:07:35 crc kubenswrapper[4848]: I1204 14:07:35.314516 4848 generic.go:334] "Generic (PLEG): container finished" podID="a5455d26-d1da-4e85-a556-a14cba35f8a9" containerID="2db9c2fb530a38acf5ff185743cbae55566e618070a496b7b76dd92b88e6c876" exitCode=0 Dec 04 14:07:35 crc kubenswrapper[4848]: I1204 14:07:35.315466 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerDied","Data":"2db9c2fb530a38acf5ff185743cbae55566e618070a496b7b76dd92b88e6c876"} Dec 04 14:07:36 crc kubenswrapper[4848]: I1204 14:07:36.352699 4848 generic.go:334] "Generic (PLEG): container finished" podID="a5455d26-d1da-4e85-a556-a14cba35f8a9" containerID="9f5667f17b1c6d90d935fe2345b3994ce369e260f797cf514217c01bb89d526d" exitCode=0 Dec 04 14:07:36 crc kubenswrapper[4848]: I1204 14:07:36.352769 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerDied","Data":"9f5667f17b1c6d90d935fe2345b3994ce369e260f797cf514217c01bb89d526d"} Dec 04 14:07:36 crc kubenswrapper[4848]: I1204 14:07:36.404057 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-pxcmt" Dec 04 14:07:37 crc kubenswrapper[4848]: I1204 14:07:37.366979 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerStarted","Data":"e96a267bb325ea8b9cda9370b6fcf00d2a55cddd3470c05af920853d7c85c36f"} Dec 04 14:07:37 crc kubenswrapper[4848]: I1204 14:07:37.367298 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerStarted","Data":"93e5bed54c941efa69cc9f3ff20681a8e3cc35733688e5c483ed518932b15d3c"} Dec 04 14:07:37 crc kubenswrapper[4848]: I1204 14:07:37.367308 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerStarted","Data":"d7d3e0d28fa11a26f299437a8cb171ebe399ce6d6648fe5118c73c3b3abb0568"} Dec 04 14:07:37 crc kubenswrapper[4848]: I1204 14:07:37.367318 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerStarted","Data":"4b552f7f6ab77857c78fd5cca3fbce9348c530f8fe1222a315da67893f1dc924"} Dec 04 14:07:37 crc kubenswrapper[4848]: I1204 14:07:37.367328 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerStarted","Data":"dda21115cd7a1ea4c78e7365cf8f3fbf8dfd91ec581df1c828530d7746bc0d75"} Dec 04 14:07:38 crc kubenswrapper[4848]: I1204 14:07:38.378583 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j45lv" event={"ID":"a5455d26-d1da-4e85-a556-a14cba35f8a9","Type":"ContainerStarted","Data":"bed445ff57ad3f7ef3bf07008e12624cab520f1c6041935cddcc111ec3916452"} Dec 04 14:07:38 crc kubenswrapper[4848]: I1204 14:07:38.378996 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:38 crc kubenswrapper[4848]: I1204 14:07:38.402960 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-j45lv" podStartSLOduration=6.821468151 podStartE2EDuration="14.402924568s" podCreationTimestamp="2025-12-04 14:07:24 +0000 UTC" firstStartedPulling="2025-12-04 14:07:24.920888428 +0000 UTC m=+1148.863384956" lastFinishedPulling="2025-12-04 14:07:32.502344815 +0000 UTC m=+1156.444841373" observedRunningTime="2025-12-04 14:07:38.399721491 +0000 UTC m=+1162.342218069" watchObservedRunningTime="2025-12-04 14:07:38.402924568 +0000 UTC m=+1162.345421096" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.596810 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-26ltw"] Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.598414 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-26ltw" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.600798 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-bnjwt" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.601663 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.602287 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.603656 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-26ltw"] Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.704789 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95tzq\" (UniqueName: \"kubernetes.io/projected/a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab-kube-api-access-95tzq\") pod \"openstack-operator-index-26ltw\" (UID: \"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab\") " pod="openstack-operators/openstack-operator-index-26ltw" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.766361 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.806384 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95tzq\" (UniqueName: \"kubernetes.io/projected/a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab-kube-api-access-95tzq\") pod \"openstack-operator-index-26ltw\" (UID: \"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab\") " pod="openstack-operators/openstack-operator-index-26ltw" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.832800 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95tzq\" (UniqueName: \"kubernetes.io/projected/a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab-kube-api-access-95tzq\") pod \"openstack-operator-index-26ltw\" (UID: \"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab\") " pod="openstack-operators/openstack-operator-index-26ltw" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.873832 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:39 crc kubenswrapper[4848]: I1204 14:07:39.917092 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-26ltw" Dec 04 14:07:40 crc kubenswrapper[4848]: I1204 14:07:40.355257 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-26ltw"] Dec 04 14:07:40 crc kubenswrapper[4848]: W1204 14:07:40.359810 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3ad8adc_cc1e_4a9b_b776_461cc2bd82ab.slice/crio-cc1cd6622dd6113236b57c37cd9789f315e30cd365de10aeb3a87e06be4e8a59 WatchSource:0}: Error finding container cc1cd6622dd6113236b57c37cd9789f315e30cd365de10aeb3a87e06be4e8a59: Status 404 returned error can't find the container with id cc1cd6622dd6113236b57c37cd9789f315e30cd365de10aeb3a87e06be4e8a59 Dec 04 14:07:40 crc kubenswrapper[4848]: I1204 14:07:40.403376 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-26ltw" event={"ID":"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab","Type":"ContainerStarted","Data":"cc1cd6622dd6113236b57c37cd9789f315e30cd365de10aeb3a87e06be4e8a59"} Dec 04 14:07:43 crc kubenswrapper[4848]: I1204 14:07:43.569396 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-26ltw"] Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.314562 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.314865 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.376202 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-st8tz"] Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.377202 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.391589 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-st8tz"] Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.400119 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkdj7\" (UniqueName: \"kubernetes.io/projected/409845c7-53dc-40ea-902b-b36b35b7d686-kube-api-access-rkdj7\") pod \"openstack-operator-index-st8tz\" (UID: \"409845c7-53dc-40ea-902b-b36b35b7d686\") " pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.434143 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-26ltw" event={"ID":"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab","Type":"ContainerStarted","Data":"084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3"} Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.434799 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-26ltw" podUID="a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab" containerName="registry-server" containerID="cri-o://084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3" gracePeriod=2 Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.502105 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkdj7\" (UniqueName: \"kubernetes.io/projected/409845c7-53dc-40ea-902b-b36b35b7d686-kube-api-access-rkdj7\") pod \"openstack-operator-index-st8tz\" (UID: \"409845c7-53dc-40ea-902b-b36b35b7d686\") " pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.543681 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkdj7\" (UniqueName: \"kubernetes.io/projected/409845c7-53dc-40ea-902b-b36b35b7d686-kube-api-access-rkdj7\") pod \"openstack-operator-index-st8tz\" (UID: \"409845c7-53dc-40ea-902b-b36b35b7d686\") " pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.704373 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.805789 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gpdjr" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.836990 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-26ltw" podStartSLOduration=2.683720447 podStartE2EDuration="5.83696735s" podCreationTimestamp="2025-12-04 14:07:39 +0000 UTC" firstStartedPulling="2025-12-04 14:07:40.362426575 +0000 UTC m=+1164.304923103" lastFinishedPulling="2025-12-04 14:07:43.515673488 +0000 UTC m=+1167.458170006" observedRunningTime="2025-12-04 14:07:44.457933359 +0000 UTC m=+1168.400429887" watchObservedRunningTime="2025-12-04 14:07:44.83696735 +0000 UTC m=+1168.779463878" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.912178 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-h8bgt" Dec 04 14:07:44 crc kubenswrapper[4848]: I1204 14:07:44.934184 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-26ltw" Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.115039 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95tzq\" (UniqueName: \"kubernetes.io/projected/a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab-kube-api-access-95tzq\") pod \"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab\" (UID: \"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab\") " Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.118464 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab-kube-api-access-95tzq" (OuterVolumeSpecName: "kube-api-access-95tzq") pod "a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab" (UID: "a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab"). InnerVolumeSpecName "kube-api-access-95tzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.202088 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-st8tz"] Dec 04 14:07:45 crc kubenswrapper[4848]: W1204 14:07:45.206615 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod409845c7_53dc_40ea_902b_b36b35b7d686.slice/crio-0f8406ce6c5ef2ad8c8d3f744dfcf546af69d69c83324f0c036df1205179332b WatchSource:0}: Error finding container 0f8406ce6c5ef2ad8c8d3f744dfcf546af69d69c83324f0c036df1205179332b: Status 404 returned error can't find the container with id 0f8406ce6c5ef2ad8c8d3f744dfcf546af69d69c83324f0c036df1205179332b Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.217382 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95tzq\" (UniqueName: \"kubernetes.io/projected/a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab-kube-api-access-95tzq\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.447921 4848 generic.go:334] "Generic (PLEG): container finished" podID="a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab" containerID="084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3" exitCode=0 Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.448004 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-26ltw" event={"ID":"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab","Type":"ContainerDied","Data":"084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3"} Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.448152 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-26ltw" Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.448405 4848 scope.go:117] "RemoveContainer" containerID="084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3" Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.448382 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-26ltw" event={"ID":"a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab","Type":"ContainerDied","Data":"cc1cd6622dd6113236b57c37cd9789f315e30cd365de10aeb3a87e06be4e8a59"} Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.450560 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-st8tz" event={"ID":"409845c7-53dc-40ea-902b-b36b35b7d686","Type":"ContainerStarted","Data":"0f8406ce6c5ef2ad8c8d3f744dfcf546af69d69c83324f0c036df1205179332b"} Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.468248 4848 scope.go:117] "RemoveContainer" containerID="084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3" Dec 04 14:07:45 crc kubenswrapper[4848]: E1204 14:07:45.468784 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3\": container with ID starting with 084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3 not found: ID does not exist" containerID="084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3" Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.468834 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3"} err="failed to get container status \"084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3\": rpc error: code = NotFound desc = could not find container \"084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3\": container with ID starting with 084d3c3f7012bf25fa3f5b5f208a495f2170d064ba9537646e50e5d72dd12eb3 not found: ID does not exist" Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.485876 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-26ltw"] Dec 04 14:07:45 crc kubenswrapper[4848]: I1204 14:07:45.494502 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-26ltw"] Dec 04 14:07:46 crc kubenswrapper[4848]: I1204 14:07:46.407199 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab" path="/var/lib/kubelet/pods/a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab/volumes" Dec 04 14:07:46 crc kubenswrapper[4848]: I1204 14:07:46.460078 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-st8tz" event={"ID":"409845c7-53dc-40ea-902b-b36b35b7d686","Type":"ContainerStarted","Data":"57070d3163a7cb498b16fc18626f68f1d41b7e1c8903fd54e29131f9d5cd77d0"} Dec 04 14:07:46 crc kubenswrapper[4848]: I1204 14:07:46.478263 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-st8tz" podStartSLOduration=2.410145379 podStartE2EDuration="2.478239227s" podCreationTimestamp="2025-12-04 14:07:44 +0000 UTC" firstStartedPulling="2025-12-04 14:07:45.210762575 +0000 UTC m=+1169.153259113" lastFinishedPulling="2025-12-04 14:07:45.278856423 +0000 UTC m=+1169.221352961" observedRunningTime="2025-12-04 14:07:46.475610913 +0000 UTC m=+1170.418107521" watchObservedRunningTime="2025-12-04 14:07:46.478239227 +0000 UTC m=+1170.420735775" Dec 04 14:07:54 crc kubenswrapper[4848]: I1204 14:07:54.705300 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:54 crc kubenswrapper[4848]: I1204 14:07:54.705904 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:54 crc kubenswrapper[4848]: I1204 14:07:54.755762 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:54 crc kubenswrapper[4848]: I1204 14:07:54.762833 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-j45lv" Dec 04 14:07:55 crc kubenswrapper[4848]: I1204 14:07:55.621202 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-st8tz" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.229061 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph"] Dec 04 14:07:58 crc kubenswrapper[4848]: E1204 14:07:58.229817 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab" containerName="registry-server" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.229838 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab" containerName="registry-server" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.230143 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3ad8adc-cc1e-4a9b-b776-461cc2bd82ab" containerName="registry-server" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.232026 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.239565 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-tcvnc" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.242146 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph"] Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.358780 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xt9m\" (UniqueName: \"kubernetes.io/projected/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-kube-api-access-7xt9m\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.358859 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-bundle\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.358964 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-util\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.460151 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-util\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.460231 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xt9m\" (UniqueName: \"kubernetes.io/projected/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-kube-api-access-7xt9m\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.460266 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-bundle\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.460719 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-bundle\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.460922 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-util\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.494390 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xt9m\" (UniqueName: \"kubernetes.io/projected/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-kube-api-access-7xt9m\") pod \"fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:58 crc kubenswrapper[4848]: I1204 14:07:58.564313 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:07:59 crc kubenswrapper[4848]: I1204 14:07:59.008001 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph"] Dec 04 14:07:59 crc kubenswrapper[4848]: W1204 14:07:59.019173 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca659f4e_033e_45a9_8d3f_898c3fdf51cb.slice/crio-f7d788cfb8884ac337dcd61f8d68214d0da2c322a544bca85a350d12553fa89f WatchSource:0}: Error finding container f7d788cfb8884ac337dcd61f8d68214d0da2c322a544bca85a350d12553fa89f: Status 404 returned error can't find the container with id f7d788cfb8884ac337dcd61f8d68214d0da2c322a544bca85a350d12553fa89f Dec 04 14:07:59 crc kubenswrapper[4848]: I1204 14:07:59.577185 4848 generic.go:334] "Generic (PLEG): container finished" podID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerID="87085ae224f338fd9e214d1bf9988a2b8520a94373b0331a50ce0407b09ae680" exitCode=0 Dec 04 14:07:59 crc kubenswrapper[4848]: I1204 14:07:59.577229 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" event={"ID":"ca659f4e-033e-45a9-8d3f-898c3fdf51cb","Type":"ContainerDied","Data":"87085ae224f338fd9e214d1bf9988a2b8520a94373b0331a50ce0407b09ae680"} Dec 04 14:07:59 crc kubenswrapper[4848]: I1204 14:07:59.577262 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" event={"ID":"ca659f4e-033e-45a9-8d3f-898c3fdf51cb","Type":"ContainerStarted","Data":"f7d788cfb8884ac337dcd61f8d68214d0da2c322a544bca85a350d12553fa89f"} Dec 04 14:08:00 crc kubenswrapper[4848]: I1204 14:08:00.588292 4848 generic.go:334] "Generic (PLEG): container finished" podID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerID="f84dcdc981b008f322bce99bfb0e522d4ac6da74a7139267a899176b0d34fc92" exitCode=0 Dec 04 14:08:00 crc kubenswrapper[4848]: I1204 14:08:00.588390 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" event={"ID":"ca659f4e-033e-45a9-8d3f-898c3fdf51cb","Type":"ContainerDied","Data":"f84dcdc981b008f322bce99bfb0e522d4ac6da74a7139267a899176b0d34fc92"} Dec 04 14:08:01 crc kubenswrapper[4848]: I1204 14:08:01.604344 4848 generic.go:334] "Generic (PLEG): container finished" podID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerID="42dbdc0c47d2e7b4d14911be35c542ae489b9d03d5cfbdfed0609de85271a6cb" exitCode=0 Dec 04 14:08:01 crc kubenswrapper[4848]: I1204 14:08:01.604408 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" event={"ID":"ca659f4e-033e-45a9-8d3f-898c3fdf51cb","Type":"ContainerDied","Data":"42dbdc0c47d2e7b4d14911be35c542ae489b9d03d5cfbdfed0609de85271a6cb"} Dec 04 14:08:02 crc kubenswrapper[4848]: I1204 14:08:02.930537 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.047035 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-bundle\") pod \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.047297 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xt9m\" (UniqueName: \"kubernetes.io/projected/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-kube-api-access-7xt9m\") pod \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.047408 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-util\") pod \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\" (UID: \"ca659f4e-033e-45a9-8d3f-898c3fdf51cb\") " Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.048416 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-bundle" (OuterVolumeSpecName: "bundle") pod "ca659f4e-033e-45a9-8d3f-898c3fdf51cb" (UID: "ca659f4e-033e-45a9-8d3f-898c3fdf51cb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.054156 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-kube-api-access-7xt9m" (OuterVolumeSpecName: "kube-api-access-7xt9m") pod "ca659f4e-033e-45a9-8d3f-898c3fdf51cb" (UID: "ca659f4e-033e-45a9-8d3f-898c3fdf51cb"). InnerVolumeSpecName "kube-api-access-7xt9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.061787 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-util" (OuterVolumeSpecName: "util") pod "ca659f4e-033e-45a9-8d3f-898c3fdf51cb" (UID: "ca659f4e-033e-45a9-8d3f-898c3fdf51cb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.149161 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xt9m\" (UniqueName: \"kubernetes.io/projected/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-kube-api-access-7xt9m\") on node \"crc\" DevicePath \"\"" Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.149191 4848 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-util\") on node \"crc\" DevicePath \"\"" Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.149203 4848 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca659f4e-033e-45a9-8d3f-898c3fdf51cb-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.638571 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" event={"ID":"ca659f4e-033e-45a9-8d3f-898c3fdf51cb","Type":"ContainerDied","Data":"f7d788cfb8884ac337dcd61f8d68214d0da2c322a544bca85a350d12553fa89f"} Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.638621 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7d788cfb8884ac337dcd61f8d68214d0da2c322a544bca85a350d12553fa89f" Dec 04 14:08:03 crc kubenswrapper[4848]: I1204 14:08:03.638663 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.665033 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt"] Dec 04 14:08:07 crc kubenswrapper[4848]: E1204 14:08:07.665834 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerName="pull" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.665849 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerName="pull" Dec 04 14:08:07 crc kubenswrapper[4848]: E1204 14:08:07.665870 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerName="util" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.665878 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerName="util" Dec 04 14:08:07 crc kubenswrapper[4848]: E1204 14:08:07.665900 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerName="extract" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.665908 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerName="extract" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.666123 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca659f4e-033e-45a9-8d3f-898c3fdf51cb" containerName="extract" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.666741 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.670223 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-fndrc" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.687252 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt"] Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.842096 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4xm8\" (UniqueName: \"kubernetes.io/projected/6641bd6c-020c-469a-8439-4953ffa85ee9-kube-api-access-c4xm8\") pod \"openstack-operator-controller-operator-799774d886-rj5dt\" (UID: \"6641bd6c-020c-469a-8439-4953ffa85ee9\") " pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.944090 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4xm8\" (UniqueName: \"kubernetes.io/projected/6641bd6c-020c-469a-8439-4953ffa85ee9-kube-api-access-c4xm8\") pod \"openstack-operator-controller-operator-799774d886-rj5dt\" (UID: \"6641bd6c-020c-469a-8439-4953ffa85ee9\") " pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.971892 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4xm8\" (UniqueName: \"kubernetes.io/projected/6641bd6c-020c-469a-8439-4953ffa85ee9-kube-api-access-c4xm8\") pod \"openstack-operator-controller-operator-799774d886-rj5dt\" (UID: \"6641bd6c-020c-469a-8439-4953ffa85ee9\") " pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" Dec 04 14:08:07 crc kubenswrapper[4848]: I1204 14:08:07.988132 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" Dec 04 14:08:08 crc kubenswrapper[4848]: I1204 14:08:08.521739 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt"] Dec 04 14:08:08 crc kubenswrapper[4848]: I1204 14:08:08.684108 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" event={"ID":"6641bd6c-020c-469a-8439-4953ffa85ee9","Type":"ContainerStarted","Data":"f48d7b35662576b4f72e6620ba9bf022a91adc794247245cd3d43131c7c59c23"} Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.313897 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.314540 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.314647 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.315567 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"131123771f065baa008e65d389e63e335b142b4c147179bfa07ab2fb29d26eae"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.315651 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://131123771f065baa008e65d389e63e335b142b4c147179bfa07ab2fb29d26eae" gracePeriod=600 Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.736780 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="131123771f065baa008e65d389e63e335b142b4c147179bfa07ab2fb29d26eae" exitCode=0 Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.736860 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"131123771f065baa008e65d389e63e335b142b4c147179bfa07ab2fb29d26eae"} Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.736906 4848 scope.go:117] "RemoveContainer" containerID="f4368b11f240a04bf6f6e83cbceaec0f4a992be59424fb78d630b078c54c789d" Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.739885 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" event={"ID":"6641bd6c-020c-469a-8439-4953ffa85ee9","Type":"ContainerStarted","Data":"49010c5733f861af240f78a63bee39be78f51c60b3cdc20a78e1b95b2998f6b0"} Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.740104 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" Dec 04 14:08:14 crc kubenswrapper[4848]: I1204 14:08:14.777009 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" podStartSLOduration=2.804205681 podStartE2EDuration="7.776987274s" podCreationTimestamp="2025-12-04 14:08:07 +0000 UTC" firstStartedPulling="2025-12-04 14:08:08.533895812 +0000 UTC m=+1192.476392330" lastFinishedPulling="2025-12-04 14:08:13.506677355 +0000 UTC m=+1197.449173923" observedRunningTime="2025-12-04 14:08:14.768538229 +0000 UTC m=+1198.711034777" watchObservedRunningTime="2025-12-04 14:08:14.776987274 +0000 UTC m=+1198.719483812" Dec 04 14:08:15 crc kubenswrapper[4848]: I1204 14:08:15.748843 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"31d6948afbef92b12725ad2b8756f5336adda03c65024c1864ef7ff8a0de5eb8"} Dec 04 14:08:27 crc kubenswrapper[4848]: I1204 14:08:27.990147 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-799774d886-rj5dt" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.796221 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.802146 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.809049 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hch8f" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.829000 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.835569 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.842923 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-zs9gf" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.858632 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.872085 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.880056 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.881719 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.886844 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-6nsr6" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.915423 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.931115 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.934257 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.940022 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-4zlrx" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.942997 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.944781 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cmsj\" (UniqueName: \"kubernetes.io/projected/038e3294-a0db-4c23-8ef2-ed6a7e617a3f-kube-api-access-7cmsj\") pod \"cinder-operator-controller-manager-859b6ccc6-9kj8j\" (UID: \"038e3294-a0db-4c23-8ef2-ed6a7e617a3f\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.944845 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5466\" (UniqueName: \"kubernetes.io/projected/4ad79d92-0097-4c46-930a-75489fad9915-kube-api-access-s5466\") pod \"barbican-operator-controller-manager-7d9dfd778-65bvj\" (UID: \"4ad79d92-0097-4c46-930a-75489fad9915\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.948279 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.952029 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-j6wzd" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.961220 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.967164 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.979020 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.980473 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.991400 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-dl6tq" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.991579 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2"] Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.992803 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.995419 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.995476 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kzvck" Dec 04 14:08:46 crc kubenswrapper[4848]: I1204 14:08:46.997962 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.042837 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.045798 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq74m\" (UniqueName: \"kubernetes.io/projected/7a7135b8-89ed-4a8a-a09e-7e343bcd30ba-kube-api-access-qq74m\") pod \"designate-operator-controller-manager-78b4bc895b-pv6z2\" (UID: \"7a7135b8-89ed-4a8a-a09e-7e343bcd30ba\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.045922 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntd5s\" (UniqueName: \"kubernetes.io/projected/7b00aa45-21c2-4be2-b702-8f191fc12b24-kube-api-access-ntd5s\") pod \"glance-operator-controller-manager-77987cd8cd-zljs9\" (UID: \"7b00aa45-21c2-4be2-b702-8f191fc12b24\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.046043 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc8fk\" (UniqueName: \"kubernetes.io/projected/1db7d065-38b0-4ef5-a468-660c4073f7f6-kube-api-access-hc8fk\") pod \"heat-operator-controller-manager-5f64f6f8bb-4zcdk\" (UID: \"1db7d065-38b0-4ef5-a468-660c4073f7f6\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.046223 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cmsj\" (UniqueName: \"kubernetes.io/projected/038e3294-a0db-4c23-8ef2-ed6a7e617a3f-kube-api-access-7cmsj\") pod \"cinder-operator-controller-manager-859b6ccc6-9kj8j\" (UID: \"038e3294-a0db-4c23-8ef2-ed6a7e617a3f\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.046332 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5466\" (UniqueName: \"kubernetes.io/projected/4ad79d92-0097-4c46-930a-75489fad9915-kube-api-access-s5466\") pod \"barbican-operator-controller-manager-7d9dfd778-65bvj\" (UID: \"4ad79d92-0097-4c46-930a-75489fad9915\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.050810 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.057797 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.068575 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-94v7g" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.069866 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.071385 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.077335 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5466\" (UniqueName: \"kubernetes.io/projected/4ad79d92-0097-4c46-930a-75489fad9915-kube-api-access-s5466\") pod \"barbican-operator-controller-manager-7d9dfd778-65bvj\" (UID: \"4ad79d92-0097-4c46-930a-75489fad9915\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.077735 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-f9wdq" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.083484 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.088411 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cmsj\" (UniqueName: \"kubernetes.io/projected/038e3294-a0db-4c23-8ef2-ed6a7e617a3f-kube-api-access-7cmsj\") pod \"cinder-operator-controller-manager-859b6ccc6-9kj8j\" (UID: \"038e3294-a0db-4c23-8ef2-ed6a7e617a3f\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.121067 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.122521 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.150863 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69mk2\" (UniqueName: \"kubernetes.io/projected/a9c2feb0-4b79-4adc-b7a6-885a5a33c244-kube-api-access-69mk2\") pod \"horizon-operator-controller-manager-68c6d99b8f-pkb5l\" (UID: \"a9c2feb0-4b79-4adc-b7a6-885a5a33c244\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.150955 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq74m\" (UniqueName: \"kubernetes.io/projected/7a7135b8-89ed-4a8a-a09e-7e343bcd30ba-kube-api-access-qq74m\") pod \"designate-operator-controller-manager-78b4bc895b-pv6z2\" (UID: \"7a7135b8-89ed-4a8a-a09e-7e343bcd30ba\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.150991 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntd5s\" (UniqueName: \"kubernetes.io/projected/7b00aa45-21c2-4be2-b702-8f191fc12b24-kube-api-access-ntd5s\") pod \"glance-operator-controller-manager-77987cd8cd-zljs9\" (UID: \"7b00aa45-21c2-4be2-b702-8f191fc12b24\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.151016 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc8fk\" (UniqueName: \"kubernetes.io/projected/1db7d065-38b0-4ef5-a468-660c4073f7f6-kube-api-access-hc8fk\") pod \"heat-operator-controller-manager-5f64f6f8bb-4zcdk\" (UID: \"1db7d065-38b0-4ef5-a468-660c4073f7f6\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.151047 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lqv2\" (UniqueName: \"kubernetes.io/projected/c22be20e-3619-45f0-944a-eff494ac7f1a-kube-api-access-6lqv2\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.151094 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.169016 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.192228 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-925th"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.193741 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.196812 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntd5s\" (UniqueName: \"kubernetes.io/projected/7b00aa45-21c2-4be2-b702-8f191fc12b24-kube-api-access-ntd5s\") pod \"glance-operator-controller-manager-77987cd8cd-zljs9\" (UID: \"7b00aa45-21c2-4be2-b702-8f191fc12b24\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.197990 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-gjps5" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.209586 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.213702 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.219225 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq74m\" (UniqueName: \"kubernetes.io/projected/7a7135b8-89ed-4a8a-a09e-7e343bcd30ba-kube-api-access-qq74m\") pod \"designate-operator-controller-manager-78b4bc895b-pv6z2\" (UID: \"7a7135b8-89ed-4a8a-a09e-7e343bcd30ba\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.219756 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.220979 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.224186 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2ph5r" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.225812 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc8fk\" (UniqueName: \"kubernetes.io/projected/1db7d065-38b0-4ef5-a468-660c4073f7f6-kube-api-access-hc8fk\") pod \"heat-operator-controller-manager-5f64f6f8bb-4zcdk\" (UID: \"1db7d065-38b0-4ef5-a468-660c4073f7f6\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.226372 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-2rbct" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.243422 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.244821 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.249772 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9gpw4" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.251285 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.261310 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.271733 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.271791 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pschn\" (UniqueName: \"kubernetes.io/projected/9e10ea48-b397-4f33-9da2-1a82a2987a9e-kube-api-access-pschn\") pod \"nova-operator-controller-manager-697bc559fc-dvhkz\" (UID: \"9e10ea48-b397-4f33-9da2-1a82a2987a9e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.271818 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69mk2\" (UniqueName: \"kubernetes.io/projected/a9c2feb0-4b79-4adc-b7a6-885a5a33c244-kube-api-access-69mk2\") pod \"horizon-operator-controller-manager-68c6d99b8f-pkb5l\" (UID: \"a9c2feb0-4b79-4adc-b7a6-885a5a33c244\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.271854 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zh5t\" (UniqueName: \"kubernetes.io/projected/82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93-kube-api-access-4zh5t\") pod \"mariadb-operator-controller-manager-56bbcc9d85-9sqtl\" (UID: \"82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.271874 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q7wz\" (UniqueName: \"kubernetes.io/projected/16e8aa39-31cb-4051-bdcf-797b4d5792c3-kube-api-access-7q7wz\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rr887\" (UID: \"16e8aa39-31cb-4051-bdcf-797b4d5792c3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.271903 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6269\" (UniqueName: \"kubernetes.io/projected/2c849714-e889-446a-9b2c-6862cdbaf422-kube-api-access-x6269\") pod \"manila-operator-controller-manager-7c79b5df47-925th\" (UID: \"2c849714-e889-446a-9b2c-6862cdbaf422\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.271963 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2v8b\" (UniqueName: \"kubernetes.io/projected/2fc8dada-7c7b-494c-ae6d-72c2737c6f66-kube-api-access-j2v8b\") pod \"ironic-operator-controller-manager-6c548fd776-4hmhf\" (UID: \"2fc8dada-7c7b-494c-ae6d-72c2737c6f66\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.271983 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lqv2\" (UniqueName: \"kubernetes.io/projected/c22be20e-3619-45f0-944a-eff494ac7f1a-kube-api-access-6lqv2\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.272001 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f2dn\" (UniqueName: \"kubernetes.io/projected/de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a-kube-api-access-6f2dn\") pod \"keystone-operator-controller-manager-7765d96ddf-kvfn2\" (UID: \"de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" Dec 04 14:08:47 crc kubenswrapper[4848]: E1204 14:08:47.272122 4848 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:47 crc kubenswrapper[4848]: E1204 14:08:47.272165 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert podName:c22be20e-3619-45f0-944a-eff494ac7f1a nodeName:}" failed. No retries permitted until 2025-12-04 14:08:47.772147126 +0000 UTC m=+1231.714643654 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert") pod "infra-operator-controller-manager-57548d458d-7g8w2" (UID: "c22be20e-3619-45f0-944a-eff494ac7f1a") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.274660 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.274862 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-925th"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.288314 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.297437 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69mk2\" (UniqueName: \"kubernetes.io/projected/a9c2feb0-4b79-4adc-b7a6-885a5a33c244-kube-api-access-69mk2\") pod \"horizon-operator-controller-manager-68c6d99b8f-pkb5l\" (UID: \"a9c2feb0-4b79-4adc-b7a6-885a5a33c244\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.297497 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.298879 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.302873 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-6fps5" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.306523 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lqv2\" (UniqueName: \"kubernetes.io/projected/c22be20e-3619-45f0-944a-eff494ac7f1a-kube-api-access-6lqv2\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.311104 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.317046 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.341116 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.345484 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.355724 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.359915 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.360627 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nczf6" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.379450 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6269\" (UniqueName: \"kubernetes.io/projected/2c849714-e889-446a-9b2c-6862cdbaf422-kube-api-access-x6269\") pod \"manila-operator-controller-manager-7c79b5df47-925th\" (UID: \"2c849714-e889-446a-9b2c-6862cdbaf422\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.379689 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2v8b\" (UniqueName: \"kubernetes.io/projected/2fc8dada-7c7b-494c-ae6d-72c2737c6f66-kube-api-access-j2v8b\") pod \"ironic-operator-controller-manager-6c548fd776-4hmhf\" (UID: \"2fc8dada-7c7b-494c-ae6d-72c2737c6f66\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.379714 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hd5x\" (UniqueName: \"kubernetes.io/projected/a60c0a88-4168-4971-8bdd-638030697efb-kube-api-access-8hd5x\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.379752 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f2dn\" (UniqueName: \"kubernetes.io/projected/de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a-kube-api-access-6f2dn\") pod \"keystone-operator-controller-manager-7765d96ddf-kvfn2\" (UID: \"de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.379900 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmxvq\" (UniqueName: \"kubernetes.io/projected/fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7-kube-api-access-qmxvq\") pod \"octavia-operator-controller-manager-998648c74-gdsqn\" (UID: \"fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.380052 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pschn\" (UniqueName: \"kubernetes.io/projected/9e10ea48-b397-4f33-9da2-1a82a2987a9e-kube-api-access-pschn\") pod \"nova-operator-controller-manager-697bc559fc-dvhkz\" (UID: \"9e10ea48-b397-4f33-9da2-1a82a2987a9e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.380084 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.380146 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zh5t\" (UniqueName: \"kubernetes.io/projected/82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93-kube-api-access-4zh5t\") pod \"mariadb-operator-controller-manager-56bbcc9d85-9sqtl\" (UID: \"82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.380165 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q7wz\" (UniqueName: \"kubernetes.io/projected/16e8aa39-31cb-4051-bdcf-797b4d5792c3-kube-api-access-7q7wz\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rr887\" (UID: \"16e8aa39-31cb-4051-bdcf-797b4d5792c3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.423483 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zh5t\" (UniqueName: \"kubernetes.io/projected/82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93-kube-api-access-4zh5t\") pod \"mariadb-operator-controller-manager-56bbcc9d85-9sqtl\" (UID: \"82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.423544 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.425410 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.425613 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6269\" (UniqueName: \"kubernetes.io/projected/2c849714-e889-446a-9b2c-6862cdbaf422-kube-api-access-x6269\") pod \"manila-operator-controller-manager-7c79b5df47-925th\" (UID: \"2c849714-e889-446a-9b2c-6862cdbaf422\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.426000 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pschn\" (UniqueName: \"kubernetes.io/projected/9e10ea48-b397-4f33-9da2-1a82a2987a9e-kube-api-access-pschn\") pod \"nova-operator-controller-manager-697bc559fc-dvhkz\" (UID: \"9e10ea48-b397-4f33-9da2-1a82a2987a9e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.426704 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2v8b\" (UniqueName: \"kubernetes.io/projected/2fc8dada-7c7b-494c-ae6d-72c2737c6f66-kube-api-access-j2v8b\") pod \"ironic-operator-controller-manager-6c548fd776-4hmhf\" (UID: \"2fc8dada-7c7b-494c-ae6d-72c2737c6f66\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.427109 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q7wz\" (UniqueName: \"kubernetes.io/projected/16e8aa39-31cb-4051-bdcf-797b4d5792c3-kube-api-access-7q7wz\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-rr887\" (UID: \"16e8aa39-31cb-4051-bdcf-797b4d5792c3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.430509 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-846mw" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.433999 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f2dn\" (UniqueName: \"kubernetes.io/projected/de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a-kube-api-access-6f2dn\") pod \"keystone-operator-controller-manager-7765d96ddf-kvfn2\" (UID: \"de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.446247 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.448639 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.481865 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmxvq\" (UniqueName: \"kubernetes.io/projected/fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7-kube-api-access-qmxvq\") pod \"octavia-operator-controller-manager-998648c74-gdsqn\" (UID: \"fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.486748 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.487138 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hd5x\" (UniqueName: \"kubernetes.io/projected/a60c0a88-4168-4971-8bdd-638030697efb-kube-api-access-8hd5x\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:47 crc kubenswrapper[4848]: E1204 14:08:47.487930 4848 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:47 crc kubenswrapper[4848]: E1204 14:08:47.494463 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert podName:a60c0a88-4168-4971-8bdd-638030697efb nodeName:}" failed. No retries permitted until 2025-12-04 14:08:47.994438426 +0000 UTC m=+1231.936934954 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" (UID: "a60c0a88-4168-4971-8bdd-638030697efb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.505520 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.506994 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.516415 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.517933 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.520763 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmxvq\" (UniqueName: \"kubernetes.io/projected/fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7-kube-api-access-qmxvq\") pod \"octavia-operator-controller-manager-998648c74-gdsqn\" (UID: \"fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.521357 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hd5x\" (UniqueName: \"kubernetes.io/projected/a60c0a88-4168-4971-8bdd-638030697efb-kube-api-access-8hd5x\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.524582 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-lsc56" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.557361 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.559600 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.562834 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-wzsfz" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.575333 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.580146 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.591243 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j55k\" (UniqueName: \"kubernetes.io/projected/14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6-kube-api-access-4j55k\") pod \"ovn-operator-controller-manager-b6456fdb6-8hd8b\" (UID: \"14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.591312 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qlxt\" (UniqueName: \"kubernetes.io/projected/82a80725-8983-447c-8f98-d4b298ae8e7e-kube-api-access-4qlxt\") pod \"swift-operator-controller-manager-5f8c65bbfc-kdftj\" (UID: \"82a80725-8983-447c-8f98-d4b298ae8e7e\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.591374 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s59j8\" (UniqueName: \"kubernetes.io/projected/f15b3026-2d34-491c-a37f-c0053e3171ee-kube-api-access-s59j8\") pod \"placement-operator-controller-manager-78f8948974-c4j2k\" (UID: \"f15b3026-2d34-491c-a37f-c0053e3171ee\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.611887 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.617899 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.620160 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.632050 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-7v7vj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.645468 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.656570 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.656786 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.667780 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.697779 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.701170 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c4ct\" (UniqueName: \"kubernetes.io/projected/2c5d62fd-6700-4804-ba6d-4d9884b9ea02-kube-api-access-5c4ct\") pod \"telemetry-operator-controller-manager-6f78686648-gktm6\" (UID: \"2c5d62fd-6700-4804-ba6d-4d9884b9ea02\") " pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.701259 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j55k\" (UniqueName: \"kubernetes.io/projected/14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6-kube-api-access-4j55k\") pod \"ovn-operator-controller-manager-b6456fdb6-8hd8b\" (UID: \"14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.701311 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qlxt\" (UniqueName: \"kubernetes.io/projected/82a80725-8983-447c-8f98-d4b298ae8e7e-kube-api-access-4qlxt\") pod \"swift-operator-controller-manager-5f8c65bbfc-kdftj\" (UID: \"82a80725-8983-447c-8f98-d4b298ae8e7e\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.701417 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s59j8\" (UniqueName: \"kubernetes.io/projected/f15b3026-2d34-491c-a37f-c0053e3171ee-kube-api-access-s59j8\") pod \"placement-operator-controller-manager-78f8948974-c4j2k\" (UID: \"f15b3026-2d34-491c-a37f-c0053e3171ee\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.711925 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.714119 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.715274 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.720768 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-jgfxh" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.722469 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.726456 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s59j8\" (UniqueName: \"kubernetes.io/projected/f15b3026-2d34-491c-a37f-c0053e3171ee-kube-api-access-s59j8\") pod \"placement-operator-controller-manager-78f8948974-c4j2k\" (UID: \"f15b3026-2d34-491c-a37f-c0053e3171ee\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.727329 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j55k\" (UniqueName: \"kubernetes.io/projected/14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6-kube-api-access-4j55k\") pod \"ovn-operator-controller-manager-b6456fdb6-8hd8b\" (UID: \"14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.730438 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qlxt\" (UniqueName: \"kubernetes.io/projected/82a80725-8983-447c-8f98-d4b298ae8e7e-kube-api-access-4qlxt\") pod \"swift-operator-controller-manager-5f8c65bbfc-kdftj\" (UID: \"82a80725-8983-447c-8f98-d4b298ae8e7e\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.747529 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.749800 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.753693 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-8rkfh" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.771850 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.797659 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.803160 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.803257 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c4ct\" (UniqueName: \"kubernetes.io/projected/2c5d62fd-6700-4804-ba6d-4d9884b9ea02-kube-api-access-5c4ct\") pod \"telemetry-operator-controller-manager-6f78686648-gktm6\" (UID: \"2c5d62fd-6700-4804-ba6d-4d9884b9ea02\") " pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.803294 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv2gz\" (UniqueName: \"kubernetes.io/projected/12439af0-d3a3-47d8-9626-b7e357af2100-kube-api-access-wv2gz\") pod \"test-operator-controller-manager-5854674fcc-z2wqt\" (UID: \"12439af0-d3a3-47d8-9626-b7e357af2100\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.803365 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhtn6\" (UniqueName: \"kubernetes.io/projected/5382b606-8cea-4b64-98ac-950f99c2a03d-kube-api-access-vhtn6\") pod \"watcher-operator-controller-manager-769dc69bc-sr98b\" (UID: \"5382b606-8cea-4b64-98ac-950f99c2a03d\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" Dec 04 14:08:47 crc kubenswrapper[4848]: E1204 14:08:47.803520 4848 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:47 crc kubenswrapper[4848]: E1204 14:08:47.803569 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert podName:c22be20e-3619-45f0-944a-eff494ac7f1a nodeName:}" failed. No retries permitted until 2025-12-04 14:08:48.803552466 +0000 UTC m=+1232.746048994 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert") pod "infra-operator-controller-manager-57548d458d-7g8w2" (UID: "c22be20e-3619-45f0-944a-eff494ac7f1a") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.816788 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.817991 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.821463 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c4ct\" (UniqueName: \"kubernetes.io/projected/2c5d62fd-6700-4804-ba6d-4d9884b9ea02-kube-api-access-5c4ct\") pod \"telemetry-operator-controller-manager-6f78686648-gktm6\" (UID: \"2c5d62fd-6700-4804-ba6d-4d9884b9ea02\") " pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.824001 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.824097 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-lkqcp" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.824142 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.826116 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.847850 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.873642 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.875264 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.878192 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-fg89c" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.888176 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh"] Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.903565 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.904548 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv2gz\" (UniqueName: \"kubernetes.io/projected/12439af0-d3a3-47d8-9626-b7e357af2100-kube-api-access-wv2gz\") pod \"test-operator-controller-manager-5854674fcc-z2wqt\" (UID: \"12439af0-d3a3-47d8-9626-b7e357af2100\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.904636 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.904664 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxz6c\" (UniqueName: \"kubernetes.io/projected/0488ea60-3643-4e40-94cd-555ac6c93043-kube-api-access-kxz6c\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5l2qh\" (UID: \"0488ea60-3643-4e40-94cd-555ac6c93043\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.904688 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhtn6\" (UniqueName: \"kubernetes.io/projected/5382b606-8cea-4b64-98ac-950f99c2a03d-kube-api-access-vhtn6\") pod \"watcher-operator-controller-manager-769dc69bc-sr98b\" (UID: \"5382b606-8cea-4b64-98ac-950f99c2a03d\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.904730 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dtst\" (UniqueName: \"kubernetes.io/projected/12302e73-00be-4ffb-9cda-3ac6126e2f0c-kube-api-access-5dtst\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.904814 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.920050 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhtn6\" (UniqueName: \"kubernetes.io/projected/5382b606-8cea-4b64-98ac-950f99c2a03d-kube-api-access-vhtn6\") pod \"watcher-operator-controller-manager-769dc69bc-sr98b\" (UID: \"5382b606-8cea-4b64-98ac-950f99c2a03d\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.921404 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv2gz\" (UniqueName: \"kubernetes.io/projected/12439af0-d3a3-47d8-9626-b7e357af2100-kube-api-access-wv2gz\") pod \"test-operator-controller-manager-5854674fcc-z2wqt\" (UID: \"12439af0-d3a3-47d8-9626-b7e357af2100\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" Dec 04 14:08:47 crc kubenswrapper[4848]: W1204 14:08:47.937929 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod038e3294_a0db_4c23_8ef2_ed6a7e617a3f.slice/crio-cb480e89318010b55b04a0062d619d303706e5689863299965c2b54b65b39ab8 WatchSource:0}: Error finding container cb480e89318010b55b04a0062d619d303706e5689863299965c2b54b65b39ab8: Status 404 returned error can't find the container with id cb480e89318010b55b04a0062d619d303706e5689863299965c2b54b65b39ab8 Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.952650 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" Dec 04 14:08:47 crc kubenswrapper[4848]: I1204 14:08:47.959789 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj"] Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.016140 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.016273 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.016311 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxz6c\" (UniqueName: \"kubernetes.io/projected/0488ea60-3643-4e40-94cd-555ac6c93043-kube-api-access-kxz6c\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5l2qh\" (UID: \"0488ea60-3643-4e40-94cd-555ac6c93043\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.016368 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dtst\" (UniqueName: \"kubernetes.io/projected/12302e73-00be-4ffb-9cda-3ac6126e2f0c-kube-api-access-5dtst\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.016454 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.016457 4848 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.016606 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:08:48.516578741 +0000 UTC m=+1232.459075269 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "webhook-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.016515 4848 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.016757 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:08:48.516739244 +0000 UTC m=+1232.459235772 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "metrics-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.016809 4848 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.016831 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert podName:a60c0a88-4168-4971-8bdd-638030697efb nodeName:}" failed. No retries permitted until 2025-12-04 14:08:49.016824786 +0000 UTC m=+1232.959321314 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" (UID: "a60c0a88-4168-4971-8bdd-638030697efb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.020075 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j"] Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.030973 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.034155 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dtst\" (UniqueName: \"kubernetes.io/projected/12302e73-00be-4ffb-9cda-3ac6126e2f0c-kube-api-access-5dtst\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.034789 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxz6c\" (UniqueName: \"kubernetes.io/projected/0488ea60-3643-4e40-94cd-555ac6c93043-kube-api-access-kxz6c\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5l2qh\" (UID: \"0488ea60-3643-4e40-94cd-555ac6c93043\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.079805 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.081517 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" event={"ID":"038e3294-a0db-4c23-8ef2-ed6a7e617a3f","Type":"ContainerStarted","Data":"cb480e89318010b55b04a0062d619d303706e5689863299965c2b54b65b39ab8"} Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.088653 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" event={"ID":"4ad79d92-0097-4c46-930a-75489fad9915","Type":"ContainerStarted","Data":"72b0b12c09b0386b653cfaa1f98138eca6a628d3de4b916c2d1ed164b2aaab38"} Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.176083 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk"] Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.195591 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.364079 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9"] Dec 04 14:08:48 crc kubenswrapper[4848]: W1204 14:08:48.371925 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b00aa45_21c2_4be2_b702_8f191fc12b24.slice/crio-1b095e68d418174cf66f13bd7072d74fa8e893b52938434b261691f7a5e0f328 WatchSource:0}: Error finding container 1b095e68d418174cf66f13bd7072d74fa8e893b52938434b261691f7a5e0f328: Status 404 returned error can't find the container with id 1b095e68d418174cf66f13bd7072d74fa8e893b52938434b261691f7a5e0f328 Dec 04 14:08:48 crc kubenswrapper[4848]: W1204 14:08:48.379594 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9c2feb0_4b79_4adc_b7a6_885a5a33c244.slice/crio-d76a52df6abe0aff9bd549043176d167332898c49798b4740c5afc69595aac50 WatchSource:0}: Error finding container d76a52df6abe0aff9bd549043176d167332898c49798b4740c5afc69595aac50: Status 404 returned error can't find the container with id d76a52df6abe0aff9bd549043176d167332898c49798b4740c5afc69595aac50 Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.380539 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l"] Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.541046 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.541217 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.541419 4848 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.541481 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:08:49.541467853 +0000 UTC m=+1233.483964371 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "webhook-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.541927 4848 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.542106 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:08:49.542080018 +0000 UTC m=+1233.484576586 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "metrics-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.853542 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.853974 4848 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: E1204 14:08:48.854055 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert podName:c22be20e-3619-45f0-944a-eff494ac7f1a nodeName:}" failed. No retries permitted until 2025-12-04 14:08:50.854033014 +0000 UTC m=+1234.796529632 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert") pod "infra-operator-controller-manager-57548d458d-7g8w2" (UID: "c22be20e-3619-45f0-944a-eff494ac7f1a") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.858583 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl"] Dec 04 14:08:48 crc kubenswrapper[4848]: W1204 14:08:48.872658 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde3bcc75_e7c5_4dc5_ac8d_8c14d21d275a.slice/crio-10da6515de80101550e796d9e69856e6411aeb5847b176c1bf0294b4244df593 WatchSource:0}: Error finding container 10da6515de80101550e796d9e69856e6411aeb5847b176c1bf0294b4244df593: Status 404 returned error can't find the container with id 10da6515de80101550e796d9e69856e6411aeb5847b176c1bf0294b4244df593 Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.878759 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2"] Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.898987 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf"] Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.910300 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2"] Dec 04 14:08:48 crc kubenswrapper[4848]: W1204 14:08:48.915561 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a7135b8_89ed_4a8a_a09e_7e343bcd30ba.slice/crio-1c93238b0103803d7d9aa175cea2e613a22c274b803bf4710e0d07afe9c81404 WatchSource:0}: Error finding container 1c93238b0103803d7d9aa175cea2e613a22c274b803bf4710e0d07afe9c81404: Status 404 returned error can't find the container with id 1c93238b0103803d7d9aa175cea2e613a22c274b803bf4710e0d07afe9c81404 Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.916084 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887"] Dec 04 14:08:48 crc kubenswrapper[4848]: W1204 14:08:48.919547 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16e8aa39_31cb_4051_bdcf_797b4d5792c3.slice/crio-c8025aa1e38f4d6bc7aec8db52ca0a7e65fd824b2f4ab9e9bb969db1b6b2fa22 WatchSource:0}: Error finding container c8025aa1e38f4d6bc7aec8db52ca0a7e65fd824b2f4ab9e9bb969db1b6b2fa22: Status 404 returned error can't find the container with id c8025aa1e38f4d6bc7aec8db52ca0a7e65fd824b2f4ab9e9bb969db1b6b2fa22 Dec 04 14:08:48 crc kubenswrapper[4848]: I1204 14:08:48.932944 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-925th"] Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.059917 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.060229 4848 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.060287 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert podName:a60c0a88-4168-4971-8bdd-638030697efb nodeName:}" failed. No retries permitted until 2025-12-04 14:08:51.0602696 +0000 UTC m=+1235.002766128 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" (UID: "a60c0a88-4168-4971-8bdd-638030697efb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.150777 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" event={"ID":"de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a","Type":"ContainerStarted","Data":"10da6515de80101550e796d9e69856e6411aeb5847b176c1bf0294b4244df593"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.160306 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" event={"ID":"2c849714-e889-446a-9b2c-6862cdbaf422","Type":"ContainerStarted","Data":"52feb2111516173e0debaa7c933b6ac47307f52e573ba8db92a673e8b90c28f7"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.165605 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" event={"ID":"16e8aa39-31cb-4051-bdcf-797b4d5792c3","Type":"ContainerStarted","Data":"c8025aa1e38f4d6bc7aec8db52ca0a7e65fd824b2f4ab9e9bb969db1b6b2fa22"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.167541 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" event={"ID":"7a7135b8-89ed-4a8a-a09e-7e343bcd30ba","Type":"ContainerStarted","Data":"1c93238b0103803d7d9aa175cea2e613a22c274b803bf4710e0d07afe9c81404"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.170343 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" event={"ID":"2fc8dada-7c7b-494c-ae6d-72c2737c6f66","Type":"ContainerStarted","Data":"fe1f36ee072942aa17ff2537b6d402d8e6dad2ca276166d4d8fdd2907311183d"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.171235 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" event={"ID":"7b00aa45-21c2-4be2-b702-8f191fc12b24","Type":"ContainerStarted","Data":"1b095e68d418174cf66f13bd7072d74fa8e893b52938434b261691f7a5e0f328"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.174662 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" event={"ID":"82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93","Type":"ContainerStarted","Data":"ff3a77a28732cbc9806a8674fdabf748e0578511932293bf2ed85bd338414ce7"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.183631 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" event={"ID":"1db7d065-38b0-4ef5-a468-660c4073f7f6","Type":"ContainerStarted","Data":"c35686ba110efba13902a32050d5c56f17d113fd62f1dc034c2b2b82629664cd"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.185424 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" event={"ID":"a9c2feb0-4b79-4adc-b7a6-885a5a33c244","Type":"ContainerStarted","Data":"d76a52df6abe0aff9bd549043176d167332898c49798b4740c5afc69595aac50"} Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.190453 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn"] Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.195779 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj"] Dec 04 14:08:49 crc kubenswrapper[4848]: W1204 14:08:49.205564 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe7a9851_7fb6_4b99_8cc3_61cdd471e5e7.slice/crio-8156f36075967a3bf0107b0cd82f7d44c90008a2d8f70f59168599b515f9b729 WatchSource:0}: Error finding container 8156f36075967a3bf0107b0cd82f7d44c90008a2d8f70f59168599b515f9b729: Status 404 returned error can't find the container with id 8156f36075967a3bf0107b0cd82f7d44c90008a2d8f70f59168599b515f9b729 Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.205705 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k"] Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.225668 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz"] Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.262769 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b"] Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.496106 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6"] Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.545274 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b"] Dec 04 14:08:49 crc kubenswrapper[4848]: W1204 14:08:49.551541 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5382b606_8cea_4b64_98ac_950f99c2a03d.slice/crio-a11a806d45b7f4225145dc0895fd22bc05dd64a7440826347073edef4498a197 WatchSource:0}: Error finding container a11a806d45b7f4225145dc0895fd22bc05dd64a7440826347073edef4498a197: Status 404 returned error can't find the container with id a11a806d45b7f4225145dc0895fd22bc05dd64a7440826347073edef4498a197 Dec 04 14:08:49 crc kubenswrapper[4848]: W1204 14:08:49.553108 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0488ea60_3643_4e40_94cd_555ac6c93043.slice/crio-2f3d0d5a1ded243e9c06c52a0753d7b672f9032639b4d6a7a013beb15aed9e2b WatchSource:0}: Error finding container 2f3d0d5a1ded243e9c06c52a0753d7b672f9032639b4d6a7a013beb15aed9e2b: Status 404 returned error can't find the container with id 2f3d0d5a1ded243e9c06c52a0753d7b672f9032639b4d6a7a013beb15aed9e2b Dec 04 14:08:49 crc kubenswrapper[4848]: W1204 14:08:49.554628 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12439af0_d3a3_47d8_9626_b7e357af2100.slice/crio-e3e075946bcc5f06f842f1944e7f7268c06c8f9552f51143162a6d9697748361 WatchSource:0}: Error finding container e3e075946bcc5f06f842f1944e7f7268c06c8f9552f51143162a6d9697748361: Status 404 returned error can't find the container with id e3e075946bcc5f06f842f1944e7f7268c06c8f9552f51143162a6d9697748361 Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.557033 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt"] Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.559411 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wv2gz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-z2wqt_openstack-operators(12439af0-d3a3-47d8-9626-b7e357af2100): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.565620 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh"] Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.566101 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wv2gz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-z2wqt_openstack-operators(12439af0-d3a3-47d8-9626-b7e357af2100): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.567398 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" podUID="12439af0-d3a3-47d8-9626-b7e357af2100" Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.567494 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:49 crc kubenswrapper[4848]: I1204 14:08:49.567653 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.567830 4848 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.567885 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:08:51.567870706 +0000 UTC m=+1235.510367234 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "webhook-server-cert" not found Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.568191 4848 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.568256 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:08:51.568221384 +0000 UTC m=+1235.510717902 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "metrics-server-cert" not found Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.568536 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vhtn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-sr98b_openstack-operators(5382b606-8cea-4b64-98ac-950f99c2a03d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.570512 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vhtn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-sr98b_openstack-operators(5382b606-8cea-4b64-98ac-950f99c2a03d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 14:08:49 crc kubenswrapper[4848]: E1204 14:08:49.571880 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" podUID="5382b606-8cea-4b64-98ac-950f99c2a03d" Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.213051 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" event={"ID":"fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7","Type":"ContainerStarted","Data":"8156f36075967a3bf0107b0cd82f7d44c90008a2d8f70f59168599b515f9b729"} Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.215021 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" event={"ID":"82a80725-8983-447c-8f98-d4b298ae8e7e","Type":"ContainerStarted","Data":"fb2547540ea4d112cc04f56fcdc5605a896b624dfcb5255a2f795b026dc25f12"} Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.216422 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" event={"ID":"f15b3026-2d34-491c-a37f-c0053e3171ee","Type":"ContainerStarted","Data":"99de9cdf484703b786d81c461cbfe9025f11b4472ce1dd1763bf2c3ddc6a9224"} Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.222132 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" event={"ID":"9e10ea48-b397-4f33-9da2-1a82a2987a9e","Type":"ContainerStarted","Data":"2e7bfd72f770dcc8e1c3ed00767ae398823a2b59a5e10e50dababee844585beb"} Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.223606 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" event={"ID":"12439af0-d3a3-47d8-9626-b7e357af2100","Type":"ContainerStarted","Data":"e3e075946bcc5f06f842f1944e7f7268c06c8f9552f51143162a6d9697748361"} Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.225506 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" event={"ID":"0488ea60-3643-4e40-94cd-555ac6c93043","Type":"ContainerStarted","Data":"2f3d0d5a1ded243e9c06c52a0753d7b672f9032639b4d6a7a013beb15aed9e2b"} Dec 04 14:08:50 crc kubenswrapper[4848]: E1204 14:08:50.227437 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" podUID="12439af0-d3a3-47d8-9626-b7e357af2100" Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.227679 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" event={"ID":"2c5d62fd-6700-4804-ba6d-4d9884b9ea02","Type":"ContainerStarted","Data":"64c51370c563590430a6c910d50fb05939d1cdf6b69b0c540ba72d0b7fce3c34"} Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.244577 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" event={"ID":"14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6","Type":"ContainerStarted","Data":"19e2621e0ff88773d6abef14a3233655410f1ad4bcc974df8deff0b0d8459cce"} Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.254018 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" event={"ID":"5382b606-8cea-4b64-98ac-950f99c2a03d","Type":"ContainerStarted","Data":"a11a806d45b7f4225145dc0895fd22bc05dd64a7440826347073edef4498a197"} Dec 04 14:08:50 crc kubenswrapper[4848]: E1204 14:08:50.262454 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" podUID="5382b606-8cea-4b64-98ac-950f99c2a03d" Dec 04 14:08:50 crc kubenswrapper[4848]: I1204 14:08:50.890385 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:50 crc kubenswrapper[4848]: E1204 14:08:50.890593 4848 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:50 crc kubenswrapper[4848]: E1204 14:08:50.890900 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert podName:c22be20e-3619-45f0-944a-eff494ac7f1a nodeName:}" failed. No retries permitted until 2025-12-04 14:08:54.890879064 +0000 UTC m=+1238.833375682 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert") pod "infra-operator-controller-manager-57548d458d-7g8w2" (UID: "c22be20e-3619-45f0-944a-eff494ac7f1a") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:51 crc kubenswrapper[4848]: I1204 14:08:51.095122 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:51 crc kubenswrapper[4848]: E1204 14:08:51.095350 4848 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:51 crc kubenswrapper[4848]: E1204 14:08:51.095424 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert podName:a60c0a88-4168-4971-8bdd-638030697efb nodeName:}" failed. No retries permitted until 2025-12-04 14:08:55.095406618 +0000 UTC m=+1239.037903136 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" (UID: "a60c0a88-4168-4971-8bdd-638030697efb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:51 crc kubenswrapper[4848]: E1204 14:08:51.264396 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" podUID="5382b606-8cea-4b64-98ac-950f99c2a03d" Dec 04 14:08:51 crc kubenswrapper[4848]: E1204 14:08:51.264930 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" podUID="12439af0-d3a3-47d8-9626-b7e357af2100" Dec 04 14:08:51 crc kubenswrapper[4848]: I1204 14:08:51.611929 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:51 crc kubenswrapper[4848]: I1204 14:08:51.612052 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:51 crc kubenswrapper[4848]: E1204 14:08:51.612238 4848 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:08:51 crc kubenswrapper[4848]: E1204 14:08:51.612287 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:08:55.612272578 +0000 UTC m=+1239.554769106 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "metrics-server-cert" not found Dec 04 14:08:51 crc kubenswrapper[4848]: E1204 14:08:51.612318 4848 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:08:51 crc kubenswrapper[4848]: E1204 14:08:51.612389 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:08:55.612373451 +0000 UTC m=+1239.554869969 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "webhook-server-cert" not found Dec 04 14:08:54 crc kubenswrapper[4848]: I1204 14:08:54.973820 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:08:54 crc kubenswrapper[4848]: E1204 14:08:54.973991 4848 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:54 crc kubenswrapper[4848]: E1204 14:08:54.974586 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert podName:c22be20e-3619-45f0-944a-eff494ac7f1a nodeName:}" failed. No retries permitted until 2025-12-04 14:09:02.974568096 +0000 UTC m=+1246.917064694 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert") pod "infra-operator-controller-manager-57548d458d-7g8w2" (UID: "c22be20e-3619-45f0-944a-eff494ac7f1a") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:08:55 crc kubenswrapper[4848]: I1204 14:08:55.177305 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:08:55 crc kubenswrapper[4848]: E1204 14:08:55.177533 4848 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:55 crc kubenswrapper[4848]: E1204 14:08:55.177620 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert podName:a60c0a88-4168-4971-8bdd-638030697efb nodeName:}" failed. No retries permitted until 2025-12-04 14:09:03.177596564 +0000 UTC m=+1247.120093172 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" (UID: "a60c0a88-4168-4971-8bdd-638030697efb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 14:08:55 crc kubenswrapper[4848]: I1204 14:08:55.688415 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:55 crc kubenswrapper[4848]: I1204 14:08:55.688546 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:08:55 crc kubenswrapper[4848]: E1204 14:08:55.688627 4848 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 14:08:55 crc kubenswrapper[4848]: E1204 14:08:55.688660 4848 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 14:08:55 crc kubenswrapper[4848]: E1204 14:08:55.688688 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:09:03.688671124 +0000 UTC m=+1247.631167652 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "metrics-server-cert" not found Dec 04 14:08:55 crc kubenswrapper[4848]: E1204 14:08:55.688703 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs podName:12302e73-00be-4ffb-9cda-3ac6126e2f0c nodeName:}" failed. No retries permitted until 2025-12-04 14:09:03.688696544 +0000 UTC m=+1247.631193072 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs") pod "openstack-operator-controller-manager-576c48d787-gx8wm" (UID: "12302e73-00be-4ffb-9cda-3ac6126e2f0c") : secret "webhook-server-cert" not found Dec 04 14:09:03 crc kubenswrapper[4848]: I1204 14:09:03.053347 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:09:03 crc kubenswrapper[4848]: E1204 14:09:03.053621 4848 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 14:09:03 crc kubenswrapper[4848]: E1204 14:09:03.055020 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert podName:c22be20e-3619-45f0-944a-eff494ac7f1a nodeName:}" failed. No retries permitted until 2025-12-04 14:09:19.05494306 +0000 UTC m=+1262.997439618 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert") pod "infra-operator-controller-manager-57548d458d-7g8w2" (UID: "c22be20e-3619-45f0-944a-eff494ac7f1a") : secret "infra-operator-webhook-server-cert" not found Dec 04 14:09:03 crc kubenswrapper[4848]: I1204 14:09:03.258471 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:09:03 crc kubenswrapper[4848]: I1204 14:09:03.268858 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a60c0a88-4168-4971-8bdd-638030697efb-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd46t68x\" (UID: \"a60c0a88-4168-4971-8bdd-638030697efb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:09:03 crc kubenswrapper[4848]: I1204 14:09:03.352702 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:09:03 crc kubenswrapper[4848]: I1204 14:09:03.769876 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:09:03 crc kubenswrapper[4848]: I1204 14:09:03.770063 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:09:03 crc kubenswrapper[4848]: I1204 14:09:03.774608 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-metrics-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:09:03 crc kubenswrapper[4848]: I1204 14:09:03.779551 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12302e73-00be-4ffb-9cda-3ac6126e2f0c-webhook-certs\") pod \"openstack-operator-controller-manager-576c48d787-gx8wm\" (UID: \"12302e73-00be-4ffb-9cda-3ac6126e2f0c\") " pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:09:04 crc kubenswrapper[4848]: I1204 14:09:04.045434 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:09:12 crc kubenswrapper[4848]: E1204 14:09:12.625616 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 04 14:09:12 crc kubenswrapper[4848]: E1204 14:09:12.626108 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qmxvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-gdsqn_openstack-operators(fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:09:13 crc kubenswrapper[4848]: E1204 14:09:13.239646 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 04 14:09:13 crc kubenswrapper[4848]: E1204 14:09:13.240404 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4j55k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-8hd8b_openstack-operators(14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:09:15 crc kubenswrapper[4848]: E1204 14:09:15.678530 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 04 14:09:15 crc kubenswrapper[4848]: E1204 14:09:15.678725 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ntd5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-zljs9_openstack-operators(7b00aa45-21c2-4be2-b702-8f191fc12b24): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:09:16 crc kubenswrapper[4848]: E1204 14:09:16.265072 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 04 14:09:16 crc kubenswrapper[4848]: E1204 14:09:16.265428 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-69mk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-pkb5l_openstack-operators(a9c2feb0-4b79-4adc-b7a6-885a5a33c244): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:09:19 crc kubenswrapper[4848]: I1204 14:09:19.089878 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:09:19 crc kubenswrapper[4848]: I1204 14:09:19.099571 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c22be20e-3619-45f0-944a-eff494ac7f1a-cert\") pod \"infra-operator-controller-manager-57548d458d-7g8w2\" (UID: \"c22be20e-3619-45f0-944a-eff494ac7f1a\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:09:19 crc kubenswrapper[4848]: I1204 14:09:19.148463 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kzvck" Dec 04 14:09:19 crc kubenswrapper[4848]: I1204 14:09:19.155477 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:09:20 crc kubenswrapper[4848]: E1204 14:09:20.692003 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 04 14:09:20 crc kubenswrapper[4848]: E1204 14:09:20.692500 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6f2dn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-kvfn2_openstack-operators(de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:09:20 crc kubenswrapper[4848]: E1204 14:09:20.799860 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.46:5001/openstack-k8s-operators/telemetry-operator:ac43f18ccd0a7178af8f4ca91b8b98e38fdefce3" Dec 04 14:09:20 crc kubenswrapper[4848]: E1204 14:09:20.799934 4848 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.46:5001/openstack-k8s-operators/telemetry-operator:ac43f18ccd0a7178af8f4ca91b8b98e38fdefce3" Dec 04 14:09:20 crc kubenswrapper[4848]: E1204 14:09:20.800236 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.46:5001/openstack-k8s-operators/telemetry-operator:ac43f18ccd0a7178af8f4ca91b8b98e38fdefce3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5c4ct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6f78686648-gktm6_openstack-operators(2c5d62fd-6700-4804-ba6d-4d9884b9ea02): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:09:21 crc kubenswrapper[4848]: E1204 14:09:21.364641 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 04 14:09:21 crc kubenswrapper[4848]: E1204 14:09:21.364823 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pschn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-dvhkz_openstack-operators(9e10ea48-b397-4f33-9da2-1a82a2987a9e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:09:22 crc kubenswrapper[4848]: E1204 14:09:22.601111 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 04 14:09:22 crc kubenswrapper[4848]: E1204 14:09:22.601556 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kxz6c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-5l2qh_openstack-operators(0488ea60-3643-4e40-94cd-555ac6c93043): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:09:22 crc kubenswrapper[4848]: E1204 14:09:22.602808 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" podUID="0488ea60-3643-4e40-94cd-555ac6c93043" Dec 04 14:09:23 crc kubenswrapper[4848]: E1204 14:09:23.589514 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" podUID="0488ea60-3643-4e40-94cd-555ac6c93043" Dec 04 14:09:23 crc kubenswrapper[4848]: I1204 14:09:23.670767 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x"] Dec 04 14:09:23 crc kubenswrapper[4848]: W1204 14:09:23.796580 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda60c0a88_4168_4971_8bdd_638030697efb.slice/crio-b157cf7c23da0f82027bb49a6a52a3e0daa63e26ff7c5c5c19835f99cd3eda61 WatchSource:0}: Error finding container b157cf7c23da0f82027bb49a6a52a3e0daa63e26ff7c5c5c19835f99cd3eda61: Status 404 returned error can't find the container with id b157cf7c23da0f82027bb49a6a52a3e0daa63e26ff7c5c5c19835f99cd3eda61 Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.200375 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2"] Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.267757 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm"] Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.640362 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" event={"ID":"2c849714-e889-446a-9b2c-6862cdbaf422","Type":"ContainerStarted","Data":"ab22d9d84e59ffadedd94daa4bf441a80e1adbbe892c8d27190525e9ec9e4223"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.646866 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" event={"ID":"82a80725-8983-447c-8f98-d4b298ae8e7e","Type":"ContainerStarted","Data":"90c977d6b936ad525b9a5cc7ec63bd736d20dbad4988a4a569d69dec34b4ba9c"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.660018 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" event={"ID":"4ad79d92-0097-4c46-930a-75489fad9915","Type":"ContainerStarted","Data":"a84387c5820823999db156bd5b96c63a82b311a0563bcf0b3b53a68f49386ea5"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.662134 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" event={"ID":"2fc8dada-7c7b-494c-ae6d-72c2737c6f66","Type":"ContainerStarted","Data":"d5c994ffa8cc0c1418159d3d17b12eb0d56dc0945015a5d7efe98b1626751fc4"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.665400 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" event={"ID":"12302e73-00be-4ffb-9cda-3ac6126e2f0c","Type":"ContainerStarted","Data":"8bf3e92dfc30855db34ab7071e651b4cd9f0a79f54eb2aa02bef6e71039ff34b"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.671291 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" event={"ID":"a60c0a88-4168-4971-8bdd-638030697efb","Type":"ContainerStarted","Data":"b157cf7c23da0f82027bb49a6a52a3e0daa63e26ff7c5c5c19835f99cd3eda61"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.679011 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" event={"ID":"82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93","Type":"ContainerStarted","Data":"afdcfb59b90a92106b387bf8062432be0810ae0c47367becde3306ed86144190"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.681591 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" event={"ID":"16e8aa39-31cb-4051-bdcf-797b4d5792c3","Type":"ContainerStarted","Data":"11ede9d5c93caaf9f50b294c692da0e57cd082ec3cadf944d2dc17b2d4f23b01"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.693974 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" event={"ID":"1db7d065-38b0-4ef5-a468-660c4073f7f6","Type":"ContainerStarted","Data":"6b7b2eda07906c76a8bc006cf32b59752ed52b7bca9071d8c46fa6fd17a83452"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.702258 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" event={"ID":"7a7135b8-89ed-4a8a-a09e-7e343bcd30ba","Type":"ContainerStarted","Data":"daf242bd7aac8aa60642f7563672f65f0f6f61daf33f01a79b85d48e11b441ee"} Dec 04 14:09:24 crc kubenswrapper[4848]: I1204 14:09:24.703830 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" event={"ID":"c22be20e-3619-45f0-944a-eff494ac7f1a","Type":"ContainerStarted","Data":"a08f9da40afc8f6d034c31ddadb03a9e109bc856ba8343d6d2dd730d6745b0c2"} Dec 04 14:09:25 crc kubenswrapper[4848]: I1204 14:09:25.728745 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" event={"ID":"038e3294-a0db-4c23-8ef2-ed6a7e617a3f","Type":"ContainerStarted","Data":"581a3252e69c31310a3409e708c96657b5ddef5bf13d7167f207cb9588086f5e"} Dec 04 14:09:27 crc kubenswrapper[4848]: I1204 14:09:27.752049 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" event={"ID":"f15b3026-2d34-491c-a37f-c0053e3171ee","Type":"ContainerStarted","Data":"0884df082154ae5c7dacb7255c21001fa1f8f4ded01ef51611bc9a72afa00295"} Dec 04 14:09:28 crc kubenswrapper[4848]: E1204 14:09:28.665348 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" podUID="fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7" Dec 04 14:09:28 crc kubenswrapper[4848]: I1204 14:09:28.763877 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" event={"ID":"12302e73-00be-4ffb-9cda-3ac6126e2f0c","Type":"ContainerStarted","Data":"fcdde6f287241ba5b979c2d4a4725209cdec133cfd8b22e9146c279ff7ac1966"} Dec 04 14:09:28 crc kubenswrapper[4848]: I1204 14:09:28.765244 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:09:28 crc kubenswrapper[4848]: I1204 14:09:28.784828 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" event={"ID":"12439af0-d3a3-47d8-9626-b7e357af2100","Type":"ContainerStarted","Data":"07ae39afd4535d0da4c1172c91d422a2c0b10c74faaba55cf609656fcc5ba6d8"} Dec 04 14:09:28 crc kubenswrapper[4848]: I1204 14:09:28.791023 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" event={"ID":"fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7","Type":"ContainerStarted","Data":"6c1699996970620fb6814c3c0e3f5f9bf77021ad58e36ce174c53d2c275871c0"} Dec 04 14:09:28 crc kubenswrapper[4848]: I1204 14:09:28.800356 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" event={"ID":"5382b606-8cea-4b64-98ac-950f99c2a03d","Type":"ContainerStarted","Data":"c3446ffdca25cbef4bb7e0e9fec9458a1c8df87c4c0d3eed38a86c375a767ce8"} Dec 04 14:09:28 crc kubenswrapper[4848]: I1204 14:09:28.799028 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" podStartSLOduration=41.798940147 podStartE2EDuration="41.798940147s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:09:28.79366605 +0000 UTC m=+1272.736162578" watchObservedRunningTime="2025-12-04 14:09:28.798940147 +0000 UTC m=+1272.741436675" Dec 04 14:09:30 crc kubenswrapper[4848]: I1204 14:09:30.850705 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" event={"ID":"82a80725-8983-447c-8f98-d4b298ae8e7e","Type":"ContainerStarted","Data":"e6fa2f5f55c1d2e5d983e70997a5dda6dd71df64a6d50aebc3fe46cb5f483ce8"} Dec 04 14:09:30 crc kubenswrapper[4848]: I1204 14:09:30.872987 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" podStartSLOduration=4.620992199 podStartE2EDuration="43.872970118s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.214235939 +0000 UTC m=+1233.156732467" lastFinishedPulling="2025-12-04 14:09:28.466213858 +0000 UTC m=+1272.408710386" observedRunningTime="2025-12-04 14:09:30.8726535 +0000 UTC m=+1274.815150028" watchObservedRunningTime="2025-12-04 14:09:30.872970118 +0000 UTC m=+1274.815466636" Dec 04 14:09:31 crc kubenswrapper[4848]: E1204 14:09:31.236037 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" podUID="de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a" Dec 04 14:09:31 crc kubenswrapper[4848]: E1204 14:09:31.408451 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" podUID="7b00aa45-21c2-4be2-b702-8f191fc12b24" Dec 04 14:09:31 crc kubenswrapper[4848]: E1204 14:09:31.710162 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" podUID="9e10ea48-b397-4f33-9da2-1a82a2987a9e" Dec 04 14:09:31 crc kubenswrapper[4848]: I1204 14:09:31.905024 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" event={"ID":"2c849714-e889-446a-9b2c-6862cdbaf422","Type":"ContainerStarted","Data":"15b617173a367979d3d028b90269da03e958d43c7f1bdb1d178a79348fc1316a"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.906099 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.910527 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.911335 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" event={"ID":"1db7d065-38b0-4ef5-a468-660c4073f7f6","Type":"ContainerStarted","Data":"78f06bb1bd1edd5cf34de0b9637ae76ae2a7f1d9887d26d9688be535b1a45ec9"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.912053 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.919163 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.921113 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" event={"ID":"5382b606-8cea-4b64-98ac-950f99c2a03d","Type":"ContainerStarted","Data":"4cd4aca77b233cea8dcdd5891265b65330b3d797877a1caf486a25de5e26fdc1"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.921470 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.927119 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" event={"ID":"f15b3026-2d34-491c-a37f-c0053e3171ee","Type":"ContainerStarted","Data":"7afc010a905d0634e1b078e2e01cdd690394fc902b3b63343175591fd188ef44"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.927251 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.930378 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" event={"ID":"2fc8dada-7c7b-494c-ae6d-72c2737c6f66","Type":"ContainerStarted","Data":"a5a23d55052d4ce704e57629dd5e3b9b07f4f9110768ebe07c419df28738c3f7"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.932184 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.934535 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-925th" podStartSLOduration=4.371526473 podStartE2EDuration="45.934516552s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.94797076 +0000 UTC m=+1232.890467288" lastFinishedPulling="2025-12-04 14:09:30.510960839 +0000 UTC m=+1274.453457367" observedRunningTime="2025-12-04 14:09:31.930031914 +0000 UTC m=+1275.872528442" watchObservedRunningTime="2025-12-04 14:09:31.934516552 +0000 UTC m=+1275.877013080" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.940810 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.946589 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" event={"ID":"12439af0-d3a3-47d8-9626-b7e357af2100","Type":"ContainerStarted","Data":"5c329dca5daa5e356d586807393763bd4110dcb542cec532c6bf32f2116edb3d"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.947014 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.960392 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" event={"ID":"fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7","Type":"ContainerStarted","Data":"e275aafc83296293b192a08c4cd746e9fde62a1a4365764657309ddb08190b2e"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.961706 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.971773 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" podStartSLOduration=10.745123956 podStartE2EDuration="44.971746644s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.568424939 +0000 UTC m=+1233.510921457" lastFinishedPulling="2025-12-04 14:09:23.795047617 +0000 UTC m=+1267.737544145" observedRunningTime="2025-12-04 14:09:31.952325504 +0000 UTC m=+1275.894822032" watchObservedRunningTime="2025-12-04 14:09:31.971746644 +0000 UTC m=+1275.914243172" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.980395 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" event={"ID":"9e10ea48-b397-4f33-9da2-1a82a2987a9e","Type":"ContainerStarted","Data":"41cb5291c035f73d8c0f8b69d2927fb366b77b8c7c8a4363f848aadfcafb3e38"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.985912 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" podStartSLOduration=3.535335542 podStartE2EDuration="44.985887617s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.226298062 +0000 UTC m=+1233.168794590" lastFinishedPulling="2025-12-04 14:09:30.676850137 +0000 UTC m=+1274.619346665" observedRunningTime="2025-12-04 14:09:31.980272291 +0000 UTC m=+1275.922768819" watchObservedRunningTime="2025-12-04 14:09:31.985887617 +0000 UTC m=+1275.928384145" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:31.990321 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" event={"ID":"7b00aa45-21c2-4be2-b702-8f191fc12b24","Type":"ContainerStarted","Data":"1d53b1d8b97178e0b7d284e295faf0e2a2e4e63332765a78a32f026a90b042b4"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.002062 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" event={"ID":"82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93","Type":"ContainerStarted","Data":"6569321a94d403ed138d4cb3662ed14e5ab885f787aaacea0605577c15211006"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.003631 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.014199 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.038182 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" event={"ID":"16e8aa39-31cb-4051-bdcf-797b4d5792c3","Type":"ContainerStarted","Data":"df4515f8d81f64ee2b84406e5f3e0a118d182b72c872e7b2cbb240df157551c5"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.039502 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.050275 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.051235 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" event={"ID":"c22be20e-3619-45f0-944a-eff494ac7f1a","Type":"ContainerStarted","Data":"39088febfa79a5868795ade2553661124590845d10cef2709dc417c6a9c53131"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.053215 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4zcdk" podStartSLOduration=3.708707847 podStartE2EDuration="46.053199227s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.256730566 +0000 UTC m=+1232.199227094" lastFinishedPulling="2025-12-04 14:09:30.601221946 +0000 UTC m=+1274.543718474" observedRunningTime="2025-12-04 14:09:32.046672039 +0000 UTC m=+1275.989168567" watchObservedRunningTime="2025-12-04 14:09:32.053199227 +0000 UTC m=+1275.995695755" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.054388 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" event={"ID":"a60c0a88-4168-4971-8bdd-638030697efb","Type":"ContainerStarted","Data":"103ab7b96f4cfa01c09129d1614832f1a4b67dcd580fd8fea9e734a4f1315c61"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.057468 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" event={"ID":"4ad79d92-0097-4c46-930a-75489fad9915","Type":"ContainerStarted","Data":"857113a63c9b75ee143ac48162f50ec5d564784295db480780c2acbdfc1f4e8b"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.061399 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.100313 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.108068 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-4hmhf" podStartSLOduration=6.525752916 podStartE2EDuration="46.108048826s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.915820511 +0000 UTC m=+1232.858317039" lastFinishedPulling="2025-12-04 14:09:28.498116421 +0000 UTC m=+1272.440612949" observedRunningTime="2025-12-04 14:09:32.080109489 +0000 UTC m=+1276.022606017" watchObservedRunningTime="2025-12-04 14:09:32.108048826 +0000 UTC m=+1276.050545354" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.116783 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" event={"ID":"de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a","Type":"ContainerStarted","Data":"c76e0d28410da069b34cfaa98d8e55145cdb22fb25f651fa399fce3d498398c0"} Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.117865 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.137220 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-kdftj" Dec 04 14:09:32 crc kubenswrapper[4848]: E1204 14:09:32.254505 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" podUID="2c5d62fd-6700-4804-ba6d-4d9884b9ea02" Dec 04 14:09:32 crc kubenswrapper[4848]: E1204 14:09:32.258682 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" podUID="a9c2feb0-4b79-4adc-b7a6-885a5a33c244" Dec 04 14:09:32 crc kubenswrapper[4848]: E1204 14:09:32.292452 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" podUID="14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.294876 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" podStartSLOduration=10.918063976 podStartE2EDuration="45.29485417s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.559282728 +0000 UTC m=+1233.501779256" lastFinishedPulling="2025-12-04 14:09:23.936072922 +0000 UTC m=+1267.878569450" observedRunningTime="2025-12-04 14:09:32.290074155 +0000 UTC m=+1276.232570693" watchObservedRunningTime="2025-12-04 14:09:32.29485417 +0000 UTC m=+1276.237350698" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.345534 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" podStartSLOduration=3.895788683 podStartE2EDuration="45.345517538s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.214136507 +0000 UTC m=+1233.156633035" lastFinishedPulling="2025-12-04 14:09:30.663865362 +0000 UTC m=+1274.606361890" observedRunningTime="2025-12-04 14:09:32.33529525 +0000 UTC m=+1276.277791768" watchObservedRunningTime="2025-12-04 14:09:32.345517538 +0000 UTC m=+1276.288014056" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.379076 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-rr887" podStartSLOduration=3.820109459 podStartE2EDuration="45.37905473s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.922286848 +0000 UTC m=+1232.864783376" lastFinishedPulling="2025-12-04 14:09:30.481232109 +0000 UTC m=+1274.423728647" observedRunningTime="2025-12-04 14:09:32.360799238 +0000 UTC m=+1276.303295766" watchObservedRunningTime="2025-12-04 14:09:32.37905473 +0000 UTC m=+1276.321551258" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.427182 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-9sqtl" podStartSLOduration=5.674133337 podStartE2EDuration="46.427163996s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.882614367 +0000 UTC m=+1232.825110895" lastFinishedPulling="2025-12-04 14:09:29.635645036 +0000 UTC m=+1273.578141554" observedRunningTime="2025-12-04 14:09:32.389308849 +0000 UTC m=+1276.331805377" watchObservedRunningTime="2025-12-04 14:09:32.427163996 +0000 UTC m=+1276.369660524" Dec 04 14:09:32 crc kubenswrapper[4848]: I1204 14:09:32.509627 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-65bvj" podStartSLOduration=3.89149841 podStartE2EDuration="46.509610263s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:47.925919977 +0000 UTC m=+1231.868416505" lastFinishedPulling="2025-12-04 14:09:30.54403183 +0000 UTC m=+1274.486528358" observedRunningTime="2025-12-04 14:09:32.509311065 +0000 UTC m=+1276.451807593" watchObservedRunningTime="2025-12-04 14:09:32.509610263 +0000 UTC m=+1276.452106781" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.125381 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" event={"ID":"c22be20e-3619-45f0-944a-eff494ac7f1a","Type":"ContainerStarted","Data":"43e8d9e1a21ec4e4a5a8d59c5c89366e4276372f599e084c65944a79b67dea2e"} Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.125689 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.128825 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" event={"ID":"a60c0a88-4168-4971-8bdd-638030697efb","Type":"ContainerStarted","Data":"3a17e2b86051689bfa664976cc03d89ca7ca85ac1e456263ac0aff69d6308ccf"} Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.129201 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.130756 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" event={"ID":"038e3294-a0db-4c23-8ef2-ed6a7e617a3f","Type":"ContainerStarted","Data":"186bb36b71125114d913665ea333337ecd054496075dca2cb3174acb06227dbe"} Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.131602 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.133142 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.133302 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" event={"ID":"a9c2feb0-4b79-4adc-b7a6-885a5a33c244","Type":"ContainerStarted","Data":"a155d0d7511335ad1bc2eca14d98184cca7707133e691c2ad1de3f6fdc76efc1"} Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.134771 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" event={"ID":"2c5d62fd-6700-4804-ba6d-4d9884b9ea02","Type":"ContainerStarted","Data":"32343e53b47932a37e3de4a3d27d66243b65260d3e9f174eae4f9bbb78c05b64"} Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.137691 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" event={"ID":"14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6","Type":"ContainerStarted","Data":"148a3708bbc5fb9e243998685eea10c983cad40781cc5f0889a94515240b375b"} Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.139874 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" event={"ID":"7a7135b8-89ed-4a8a-a09e-7e343bcd30ba","Type":"ContainerStarted","Data":"f33b38c181eed4d827cbbaaa8da63fabeb7d179c400543c0bcb62c5a754b191b"} Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.142228 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c4j2k" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.147531 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z2wqt" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.147611 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-sr98b" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.189623 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" podStartSLOduration=39.502067859 podStartE2EDuration="46.189603234s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:09:23.800525849 +0000 UTC m=+1267.743022377" lastFinishedPulling="2025-12-04 14:09:30.488061224 +0000 UTC m=+1274.430557752" observedRunningTime="2025-12-04 14:09:33.188972999 +0000 UTC m=+1277.131469537" watchObservedRunningTime="2025-12-04 14:09:33.189603234 +0000 UTC m=+1277.132099762" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.194598 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" podStartSLOduration=40.782311689 podStartE2EDuration="47.194587315s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:09:24.243376077 +0000 UTC m=+1268.185872605" lastFinishedPulling="2025-12-04 14:09:30.655651703 +0000 UTC m=+1274.598148231" observedRunningTime="2025-12-04 14:09:33.146791658 +0000 UTC m=+1277.089288206" watchObservedRunningTime="2025-12-04 14:09:33.194587315 +0000 UTC m=+1277.137083843" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.314151 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" podStartSLOduration=5.634457276 podStartE2EDuration="47.314133221s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.919956761 +0000 UTC m=+1232.862453289" lastFinishedPulling="2025-12-04 14:09:30.599632706 +0000 UTC m=+1274.542129234" observedRunningTime="2025-12-04 14:09:33.306032614 +0000 UTC m=+1277.248529142" watchObservedRunningTime="2025-12-04 14:09:33.314133221 +0000 UTC m=+1277.256629739" Dec 04 14:09:33 crc kubenswrapper[4848]: I1204 14:09:33.368191 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-9kj8j" podStartSLOduration=4.521821547 podStartE2EDuration="47.36817455s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:47.943864481 +0000 UTC m=+1231.886361009" lastFinishedPulling="2025-12-04 14:09:30.790217484 +0000 UTC m=+1274.732714012" observedRunningTime="2025-12-04 14:09:33.345242904 +0000 UTC m=+1277.287739432" watchObservedRunningTime="2025-12-04 14:09:33.36817455 +0000 UTC m=+1277.310671068" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.067587 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.223886 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" event={"ID":"de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a","Type":"ContainerStarted","Data":"0ff4ee3bade7a781519a90fa804c05f3e3a565703546e0f089c94343c3ac26c8"} Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.229606 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.229651 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" event={"ID":"7b00aa45-21c2-4be2-b702-8f191fc12b24","Type":"ContainerStarted","Data":"0accc4e164b641cf7782126f093854f6223e0701e3b70d243ab5ea8f87118af2"} Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.229793 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.243723 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" podStartSLOduration=4.426215218 podStartE2EDuration="48.243706258s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.880365543 +0000 UTC m=+1232.822862071" lastFinishedPulling="2025-12-04 14:09:32.697856583 +0000 UTC m=+1276.640353111" observedRunningTime="2025-12-04 14:09:34.240782198 +0000 UTC m=+1278.183278726" watchObservedRunningTime="2025-12-04 14:09:34.243706258 +0000 UTC m=+1278.186202786" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.245160 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.258063 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" podStartSLOduration=4.064671741 podStartE2EDuration="48.258046506s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.379713385 +0000 UTC m=+1232.322209913" lastFinishedPulling="2025-12-04 14:09:32.57308815 +0000 UTC m=+1276.515584678" observedRunningTime="2025-12-04 14:09:34.256059578 +0000 UTC m=+1278.198556106" watchObservedRunningTime="2025-12-04 14:09:34.258046506 +0000 UTC m=+1278.200543034" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.279711 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" event={"ID":"9e10ea48-b397-4f33-9da2-1a82a2987a9e","Type":"ContainerStarted","Data":"61900788185ac38e2be1af1705384d0b3aea4ec993fafeeacb49f0debfbb64f7"} Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.279978 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.288362 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-pv6z2" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.301683 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" podStartSLOduration=2.866247354 podStartE2EDuration="47.301659903s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.518044589 +0000 UTC m=+1233.460541117" lastFinishedPulling="2025-12-04 14:09:33.953457138 +0000 UTC m=+1277.895953666" observedRunningTime="2025-12-04 14:09:34.285055721 +0000 UTC m=+1278.227552249" watchObservedRunningTime="2025-12-04 14:09:34.301659903 +0000 UTC m=+1278.244156431" Dec 04 14:09:34 crc kubenswrapper[4848]: I1204 14:09:34.318551 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" podStartSLOduration=3.977949774 podStartE2EDuration="47.318534222s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.237449032 +0000 UTC m=+1233.179945560" lastFinishedPulling="2025-12-04 14:09:32.57803349 +0000 UTC m=+1276.520530008" observedRunningTime="2025-12-04 14:09:34.316288207 +0000 UTC m=+1278.258784735" watchObservedRunningTime="2025-12-04 14:09:34.318534222 +0000 UTC m=+1278.261030750" Dec 04 14:09:35 crc kubenswrapper[4848]: I1204 14:09:35.289406 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" event={"ID":"a9c2feb0-4b79-4adc-b7a6-885a5a33c244","Type":"ContainerStarted","Data":"94b487d2aea61ab768c9a07cad12f3bdb5b2fa5a42231a1d67bc69bb83922fd9"} Dec 04 14:09:35 crc kubenswrapper[4848]: I1204 14:09:35.289728 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" Dec 04 14:09:35 crc kubenswrapper[4848]: I1204 14:09:35.292062 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" event={"ID":"2c5d62fd-6700-4804-ba6d-4d9884b9ea02","Type":"ContainerStarted","Data":"5f3cd23e3226d5cf2a985523b56b47894ab4503bc5a46007572bd41735177a3f"} Dec 04 14:09:35 crc kubenswrapper[4848]: I1204 14:09:35.294099 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" event={"ID":"14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6","Type":"ContainerStarted","Data":"3c69176b07b617846af2c118e0e93b4706c56e7e2d7c426ac5d0509bcf8c03cc"} Dec 04 14:09:35 crc kubenswrapper[4848]: I1204 14:09:35.294585 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" Dec 04 14:09:35 crc kubenswrapper[4848]: I1204 14:09:35.314277 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" podStartSLOduration=3.5415532389999997 podStartE2EDuration="49.314260821s" podCreationTimestamp="2025-12-04 14:08:46 +0000 UTC" firstStartedPulling="2025-12-04 14:08:48.384534472 +0000 UTC m=+1232.327031000" lastFinishedPulling="2025-12-04 14:09:34.157242054 +0000 UTC m=+1278.099738582" observedRunningTime="2025-12-04 14:09:35.308104312 +0000 UTC m=+1279.250600860" watchObservedRunningTime="2025-12-04 14:09:35.314260821 +0000 UTC m=+1279.256757349" Dec 04 14:09:35 crc kubenswrapper[4848]: I1204 14:09:35.331237 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" podStartSLOduration=3.393310562 podStartE2EDuration="48.331215382s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.226233401 +0000 UTC m=+1233.168729929" lastFinishedPulling="2025-12-04 14:09:34.164138221 +0000 UTC m=+1278.106634749" observedRunningTime="2025-12-04 14:09:35.324997631 +0000 UTC m=+1279.267494199" watchObservedRunningTime="2025-12-04 14:09:35.331215382 +0000 UTC m=+1279.273711920" Dec 04 14:09:36 crc kubenswrapper[4848]: I1204 14:09:36.301807 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" Dec 04 14:09:37 crc kubenswrapper[4848]: I1204 14:09:37.724985 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-gdsqn" Dec 04 14:09:39 crc kubenswrapper[4848]: I1204 14:09:39.165052 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-7g8w2" Dec 04 14:09:39 crc kubenswrapper[4848]: I1204 14:09:39.326128 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" event={"ID":"0488ea60-3643-4e40-94cd-555ac6c93043","Type":"ContainerStarted","Data":"ed58fab8a6f292354da94d117055a91dbad50b4678b290e4c30e81f6d0079529"} Dec 04 14:09:39 crc kubenswrapper[4848]: I1204 14:09:39.374233 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5l2qh" podStartSLOduration=3.006760897 podStartE2EDuration="52.374212046s" podCreationTimestamp="2025-12-04 14:08:47 +0000 UTC" firstStartedPulling="2025-12-04 14:08:49.5564586 +0000 UTC m=+1233.498955128" lastFinishedPulling="2025-12-04 14:09:38.923909739 +0000 UTC m=+1282.866406277" observedRunningTime="2025-12-04 14:09:39.357914942 +0000 UTC m=+1283.300411460" watchObservedRunningTime="2025-12-04 14:09:39.374212046 +0000 UTC m=+1283.316708574" Dec 04 14:09:43 crc kubenswrapper[4848]: I1204 14:09:43.366478 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" Dec 04 14:09:47 crc kubenswrapper[4848]: I1204 14:09:47.265679 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-zljs9" Dec 04 14:09:47 crc kubenswrapper[4848]: I1204 14:09:47.319064 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pkb5l" Dec 04 14:09:47 crc kubenswrapper[4848]: I1204 14:09:47.578191 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-kvfn2" Dec 04 14:09:47 crc kubenswrapper[4848]: I1204 14:09:47.701409 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-dvhkz" Dec 04 14:09:47 crc kubenswrapper[4848]: I1204 14:09:47.778577 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-8hd8b" Dec 04 14:09:47 crc kubenswrapper[4848]: I1204 14:09:47.956766 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6f78686648-gktm6" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.884857 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwfg8"] Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.886922 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.891691 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.891757 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.891830 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.892131 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-h77hk" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.896804 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwfg8"] Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.938070 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-znvrh"] Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.943454 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.945833 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.947223 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-znvrh"] Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.979116 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02864bf6-0bad-49c3-a676-7117748f3ef6-config\") pod \"dnsmasq-dns-675f4bcbfc-nwfg8\" (UID: \"02864bf6-0bad-49c3-a676-7117748f3ef6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:02 crc kubenswrapper[4848]: I1204 14:10:02.979227 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6lx6\" (UniqueName: \"kubernetes.io/projected/02864bf6-0bad-49c3-a676-7117748f3ef6-kube-api-access-m6lx6\") pod \"dnsmasq-dns-675f4bcbfc-nwfg8\" (UID: \"02864bf6-0bad-49c3-a676-7117748f3ef6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.080916 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq7t8\" (UniqueName: \"kubernetes.io/projected/a2cf557a-089e-4c9d-b194-9ed01df93b6c-kube-api-access-sq7t8\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.080997 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02864bf6-0bad-49c3-a676-7117748f3ef6-config\") pod \"dnsmasq-dns-675f4bcbfc-nwfg8\" (UID: \"02864bf6-0bad-49c3-a676-7117748f3ef6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.081019 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-config\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.081069 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.081101 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6lx6\" (UniqueName: \"kubernetes.io/projected/02864bf6-0bad-49c3-a676-7117748f3ef6-kube-api-access-m6lx6\") pod \"dnsmasq-dns-675f4bcbfc-nwfg8\" (UID: \"02864bf6-0bad-49c3-a676-7117748f3ef6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.082201 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02864bf6-0bad-49c3-a676-7117748f3ef6-config\") pod \"dnsmasq-dns-675f4bcbfc-nwfg8\" (UID: \"02864bf6-0bad-49c3-a676-7117748f3ef6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.098983 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6lx6\" (UniqueName: \"kubernetes.io/projected/02864bf6-0bad-49c3-a676-7117748f3ef6-kube-api-access-m6lx6\") pod \"dnsmasq-dns-675f4bcbfc-nwfg8\" (UID: \"02864bf6-0bad-49c3-a676-7117748f3ef6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.182298 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq7t8\" (UniqueName: \"kubernetes.io/projected/a2cf557a-089e-4c9d-b194-9ed01df93b6c-kube-api-access-sq7t8\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.182905 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-config\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.183103 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.184479 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.185560 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-config\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.217468 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.225035 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq7t8\" (UniqueName: \"kubernetes.io/projected/a2cf557a-089e-4c9d-b194-9ed01df93b6c-kube-api-access-sq7t8\") pod \"dnsmasq-dns-78dd6ddcc-znvrh\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.281358 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.679699 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwfg8"] Dec 04 14:10:03 crc kubenswrapper[4848]: W1204 14:10:03.814576 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2cf557a_089e_4c9d_b194_9ed01df93b6c.slice/crio-c22dc559b743483e65872621de4b84d2a90a889500ab51128fd456342afe7ecd WatchSource:0}: Error finding container c22dc559b743483e65872621de4b84d2a90a889500ab51128fd456342afe7ecd: Status 404 returned error can't find the container with id c22dc559b743483e65872621de4b84d2a90a889500ab51128fd456342afe7ecd Dec 04 14:10:03 crc kubenswrapper[4848]: I1204 14:10:03.817355 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-znvrh"] Dec 04 14:10:04 crc kubenswrapper[4848]: I1204 14:10:04.567642 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" event={"ID":"02864bf6-0bad-49c3-a676-7117748f3ef6","Type":"ContainerStarted","Data":"f63a7aba2da8fd52deb26bf84a96bcb2db8ce0b6926f95c5a2b30ea60b710a9c"} Dec 04 14:10:04 crc kubenswrapper[4848]: I1204 14:10:04.569672 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" event={"ID":"a2cf557a-089e-4c9d-b194-9ed01df93b6c","Type":"ContainerStarted","Data":"c22dc559b743483e65872621de4b84d2a90a889500ab51128fd456342afe7ecd"} Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.061719 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwfg8"] Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.114517 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-grcrh"] Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.116510 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.127079 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-grcrh"] Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.267972 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmszc\" (UniqueName: \"kubernetes.io/projected/c1c86a95-c98a-4049-901d-c2ac744d42f8-kube-api-access-nmszc\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.268016 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.268046 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-config\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.371191 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmszc\" (UniqueName: \"kubernetes.io/projected/c1c86a95-c98a-4049-901d-c2ac744d42f8-kube-api-access-nmszc\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.371235 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.371265 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-config\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.372399 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.372455 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-config\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.405047 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmszc\" (UniqueName: \"kubernetes.io/projected/c1c86a95-c98a-4049-901d-c2ac744d42f8-kube-api-access-nmszc\") pod \"dnsmasq-dns-5ccc8479f9-grcrh\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.439602 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.444759 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-znvrh"] Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.462981 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dpbbw"] Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.465061 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.484881 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dpbbw"] Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.578056 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x6qs\" (UniqueName: \"kubernetes.io/projected/f66fc099-8a2e-49ea-803c-417aeedbe389-kube-api-access-2x6qs\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.578099 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-config\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.578168 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.680612 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x6qs\" (UniqueName: \"kubernetes.io/projected/f66fc099-8a2e-49ea-803c-417aeedbe389-kube-api-access-2x6qs\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.680655 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-config\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.680717 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.681981 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-config\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.682350 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.707842 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x6qs\" (UniqueName: \"kubernetes.io/projected/f66fc099-8a2e-49ea-803c-417aeedbe389-kube-api-access-2x6qs\") pod \"dnsmasq-dns-57d769cc4f-dpbbw\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:06 crc kubenswrapper[4848]: I1204 14:10:06.831539 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.103168 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-grcrh"] Dec 04 14:10:07 crc kubenswrapper[4848]: W1204 14:10:07.123694 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1c86a95_c98a_4049_901d_c2ac744d42f8.slice/crio-5dc87611d6e63349296355dfa51b2e14b790d6e5d1b8c5513f334288d58e5f73 WatchSource:0}: Error finding container 5dc87611d6e63349296355dfa51b2e14b790d6e5d1b8c5513f334288d58e5f73: Status 404 returned error can't find the container with id 5dc87611d6e63349296355dfa51b2e14b790d6e5d1b8c5513f334288d58e5f73 Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.276431 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.277995 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.279974 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xdqv6" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.280253 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.281211 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.281283 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.281510 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.281508 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.282245 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.285434 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.331042 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dpbbw"] Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408036 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408265 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408393 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30883c46-53ec-45f9-84f6-29a64ccf8994-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408480 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408606 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30883c46-53ec-45f9-84f6-29a64ccf8994-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408711 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408796 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408859 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.408969 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.409076 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.409251 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kss47\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-kube-api-access-kss47\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510444 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30883c46-53ec-45f9-84f6-29a64ccf8994-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510491 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510518 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30883c46-53ec-45f9-84f6-29a64ccf8994-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510537 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510563 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510579 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510609 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510640 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510677 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kss47\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-kube-api-access-kss47\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510700 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.510724 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.511144 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.511285 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.511377 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.511701 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.512327 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.513182 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.522617 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30883c46-53ec-45f9-84f6-29a64ccf8994-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.523677 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.526564 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30883c46-53ec-45f9-84f6-29a64ccf8994-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.531660 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kss47\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-kube-api-access-kss47\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.533587 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.536919 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.599162 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.601863 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.603642 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.606700 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.608666 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.608816 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.611539 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.611917 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-lv446" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.612131 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.612367 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.619938 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.672246 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" event={"ID":"f66fc099-8a2e-49ea-803c-417aeedbe389","Type":"ContainerStarted","Data":"b0fb18610625685b47feb11a9c646e7e48136a5356be9218e7b8bfd710809bc3"} Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.673441 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" event={"ID":"c1c86a95-c98a-4049-901d-c2ac744d42f8","Type":"ContainerStarted","Data":"5dc87611d6e63349296355dfa51b2e14b790d6e5d1b8c5513f334288d58e5f73"} Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715137 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715181 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zg8j\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-kube-api-access-8zg8j\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715226 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715271 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6049ed43-40db-4bff-b89b-88cce285fd2d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715302 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-config-data\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715319 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715339 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715363 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715412 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715701 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6049ed43-40db-4bff-b89b-88cce285fd2d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.715782 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817340 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-config-data\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817691 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817713 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817735 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817760 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817821 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6049ed43-40db-4bff-b89b-88cce285fd2d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817853 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817876 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817893 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zg8j\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-kube-api-access-8zg8j\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.817931 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.818004 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6049ed43-40db-4bff-b89b-88cce285fd2d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.818408 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.818712 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.819306 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.819400 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.819888 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-config-data\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.820007 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.826831 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.831821 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6049ed43-40db-4bff-b89b-88cce285fd2d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.836397 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6049ed43-40db-4bff-b89b-88cce285fd2d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.837636 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.841556 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zg8j\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-kube-api-access-8zg8j\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.846606 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " pod="openstack/rabbitmq-server-0" Dec 04 14:10:07 crc kubenswrapper[4848]: I1204 14:10:07.924122 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:10:08 crc kubenswrapper[4848]: I1204 14:10:08.086278 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.042044 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.045183 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.053014 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-snqcv" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.053867 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.054216 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.054351 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.060471 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.065595 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.144509 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-config-data-default\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.144556 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.144600 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.144668 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.144736 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-kolla-config\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.144761 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.144779 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.144810 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c27xz\" (UniqueName: \"kubernetes.io/projected/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-kube-api-access-c27xz\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.246930 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-kolla-config\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.247065 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.247185 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.247232 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c27xz\" (UniqueName: \"kubernetes.io/projected/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-kube-api-access-c27xz\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.247332 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-config-data-default\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.247361 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.247682 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.247897 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-kolla-config\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.248493 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-config-data-default\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.248708 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.248773 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.251548 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.251907 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.256591 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.265588 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.270806 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c27xz\" (UniqueName: \"kubernetes.io/projected/f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52-kube-api-access-c27xz\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.290286 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52\") " pod="openstack/openstack-galera-0" Dec 04 14:10:09 crc kubenswrapper[4848]: I1204 14:10:09.386480 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.478864 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.481272 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.483736 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-mmjzg" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.484072 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.484246 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.486984 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.490917 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.570907 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.571029 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.571094 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.571129 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.571156 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxff5\" (UniqueName: \"kubernetes.io/projected/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-kube-api-access-gxff5\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.571192 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.571258 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.571292 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.665777 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.667141 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.670271 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.670514 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nm5rz" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.670670 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.672763 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.672819 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.673066 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.673089 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.673115 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxff5\" (UniqueName: \"kubernetes.io/projected/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-kube-api-access-gxff5\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.673139 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.673179 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.673202 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.673846 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.674137 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.673908 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.676658 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.681503 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.690605 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.700103 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.705541 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.722243 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxff5\" (UniqueName: \"kubernetes.io/projected/1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2-kube-api-access-gxff5\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.734735 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2\") " pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.776125 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/add25985-bc0c-48ee-9f42-4404503638b5-kolla-config\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.776186 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add25985-bc0c-48ee-9f42-4404503638b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.776330 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/add25985-bc0c-48ee-9f42-4404503638b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.776365 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/add25985-bc0c-48ee-9f42-4404503638b5-config-data\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.776409 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slzdx\" (UniqueName: \"kubernetes.io/projected/add25985-bc0c-48ee-9f42-4404503638b5-kube-api-access-slzdx\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.862463 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.877797 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/add25985-bc0c-48ee-9f42-4404503638b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.877849 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/add25985-bc0c-48ee-9f42-4404503638b5-config-data\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.877881 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slzdx\" (UniqueName: \"kubernetes.io/projected/add25985-bc0c-48ee-9f42-4404503638b5-kube-api-access-slzdx\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.877921 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/add25985-bc0c-48ee-9f42-4404503638b5-kolla-config\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.877943 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add25985-bc0c-48ee-9f42-4404503638b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.878912 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/add25985-bc0c-48ee-9f42-4404503638b5-kolla-config\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.878978 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/add25985-bc0c-48ee-9f42-4404503638b5-config-data\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.881074 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/add25985-bc0c-48ee-9f42-4404503638b5-combined-ca-bundle\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.883135 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/add25985-bc0c-48ee-9f42-4404503638b5-memcached-tls-certs\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:10 crc kubenswrapper[4848]: I1204 14:10:10.903942 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slzdx\" (UniqueName: \"kubernetes.io/projected/add25985-bc0c-48ee-9f42-4404503638b5-kube-api-access-slzdx\") pod \"memcached-0\" (UID: \"add25985-bc0c-48ee-9f42-4404503638b5\") " pod="openstack/memcached-0" Dec 04 14:10:11 crc kubenswrapper[4848]: I1204 14:10:11.093521 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 14:10:11 crc kubenswrapper[4848]: W1204 14:10:11.767988 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30883c46_53ec_45f9_84f6_29a64ccf8994.slice/crio-f0f6cd6caa4cc643f3b4098b4205579b53c6ef21e612987957414c0e257cc923 WatchSource:0}: Error finding container f0f6cd6caa4cc643f3b4098b4205579b53c6ef21e612987957414c0e257cc923: Status 404 returned error can't find the container with id f0f6cd6caa4cc643f3b4098b4205579b53c6ef21e612987957414c0e257cc923 Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.571037 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.572177 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.576772 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-f2mjv" Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.587134 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.622087 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zz5x\" (UniqueName: \"kubernetes.io/projected/d644fdb6-4d43-4e46-9537-f2fedef790e0-kube-api-access-2zz5x\") pod \"kube-state-metrics-0\" (UID: \"d644fdb6-4d43-4e46-9537-f2fedef790e0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.723870 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zz5x\" (UniqueName: \"kubernetes.io/projected/d644fdb6-4d43-4e46-9537-f2fedef790e0-kube-api-access-2zz5x\") pod \"kube-state-metrics-0\" (UID: \"d644fdb6-4d43-4e46-9537-f2fedef790e0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.747491 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30883c46-53ec-45f9-84f6-29a64ccf8994","Type":"ContainerStarted","Data":"f0f6cd6caa4cc643f3b4098b4205579b53c6ef21e612987957414c0e257cc923"} Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.755397 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zz5x\" (UniqueName: \"kubernetes.io/projected/d644fdb6-4d43-4e46-9537-f2fedef790e0-kube-api-access-2zz5x\") pod \"kube-state-metrics-0\" (UID: \"d644fdb6-4d43-4e46-9537-f2fedef790e0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:10:12 crc kubenswrapper[4848]: I1204 14:10:12.896387 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.366619 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7"] Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.373893 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.376482 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7"] Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.383314 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-4tkfg" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.383581 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.451718 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8efe0158-56d4-4526-9683-263fb1ce6f23-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-49kl7\" (UID: \"8efe0158-56d4-4526-9683-263fb1ce6f23\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.451906 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm755\" (UniqueName: \"kubernetes.io/projected/8efe0158-56d4-4526-9683-263fb1ce6f23-kube-api-access-lm755\") pod \"observability-ui-dashboards-7d5fb4cbfb-49kl7\" (UID: \"8efe0158-56d4-4526-9683-263fb1ce6f23\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.553696 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm755\" (UniqueName: \"kubernetes.io/projected/8efe0158-56d4-4526-9683-263fb1ce6f23-kube-api-access-lm755\") pod \"observability-ui-dashboards-7d5fb4cbfb-49kl7\" (UID: \"8efe0158-56d4-4526-9683-263fb1ce6f23\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.554052 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8efe0158-56d4-4526-9683-263fb1ce6f23-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-49kl7\" (UID: \"8efe0158-56d4-4526-9683-263fb1ce6f23\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:13 crc kubenswrapper[4848]: E1204 14:10:13.554191 4848 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Dec 04 14:10:13 crc kubenswrapper[4848]: E1204 14:10:13.554281 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8efe0158-56d4-4526-9683-263fb1ce6f23-serving-cert podName:8efe0158-56d4-4526-9683-263fb1ce6f23 nodeName:}" failed. No retries permitted until 2025-12-04 14:10:14.054258845 +0000 UTC m=+1317.996755373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/8efe0158-56d4-4526-9683-263fb1ce6f23-serving-cert") pod "observability-ui-dashboards-7d5fb4cbfb-49kl7" (UID: "8efe0158-56d4-4526-9683-263fb1ce6f23") : secret "observability-ui-dashboards" not found Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.585830 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm755\" (UniqueName: \"kubernetes.io/projected/8efe0158-56d4-4526-9683-263fb1ce6f23-kube-api-access-lm755\") pod \"observability-ui-dashboards-7d5fb4cbfb-49kl7\" (UID: \"8efe0158-56d4-4526-9683-263fb1ce6f23\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.721639 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5db675fff6-2jrj6"] Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.723236 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.748250 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5db675fff6-2jrj6"] Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.867196 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-service-ca\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.867782 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-oauth-serving-cert\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.867935 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-console-config\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.868075 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6167e58f-c3af-424d-9687-0c5266be601f-console-serving-cert\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.868210 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6167e58f-c3af-424d-9687-0c5266be601f-console-oauth-config\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.868318 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsj29\" (UniqueName: \"kubernetes.io/projected/6167e58f-c3af-424d-9687-0c5266be601f-kube-api-access-jsj29\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.868478 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-trusted-ca-bundle\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.970062 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-service-ca\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.970110 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-oauth-serving-cert\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.970146 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-console-config\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.970167 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6167e58f-c3af-424d-9687-0c5266be601f-console-serving-cert\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.970203 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6167e58f-c3af-424d-9687-0c5266be601f-console-oauth-config\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.970231 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsj29\" (UniqueName: \"kubernetes.io/projected/6167e58f-c3af-424d-9687-0c5266be601f-kube-api-access-jsj29\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.970289 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-trusted-ca-bundle\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.971270 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-trusted-ca-bundle\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.971770 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-service-ca\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.972062 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-console-config\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.972420 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6167e58f-c3af-424d-9687-0c5266be601f-oauth-serving-cert\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.985424 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6167e58f-c3af-424d-9687-0c5266be601f-console-serving-cert\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.988421 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsj29\" (UniqueName: \"kubernetes.io/projected/6167e58f-c3af-424d-9687-0c5266be601f-kube-api-access-jsj29\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:13 crc kubenswrapper[4848]: I1204 14:10:13.991928 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6167e58f-c3af-424d-9687-0c5266be601f-console-oauth-config\") pod \"console-5db675fff6-2jrj6\" (UID: \"6167e58f-c3af-424d-9687-0c5266be601f\") " pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.039061 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.047773 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.052065 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.061496 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.061568 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-dzcpz" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.062848 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.063986 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.064117 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.066270 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.070837 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.071851 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8efe0158-56d4-4526-9683-263fb1ce6f23-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-49kl7\" (UID: \"8efe0158-56d4-4526-9683-263fb1ce6f23\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.078647 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8efe0158-56d4-4526-9683-263fb1ce6f23-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-49kl7\" (UID: \"8efe0158-56d4-4526-9683-263fb1ce6f23\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.173641 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.174099 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cmxf\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-kube-api-access-8cmxf\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.174164 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.174236 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.174400 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.174441 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.174458 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.174532 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.276461 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cmxf\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-kube-api-access-8cmxf\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.276538 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.276598 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.276653 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.276701 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.276721 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.276768 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.276815 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.277426 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.278926 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.281290 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.281554 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.281809 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.289490 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.295912 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cmxf\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-kube-api-access-8cmxf\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.296567 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.301405 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.301725 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" Dec 04 14:10:14 crc kubenswrapper[4848]: I1204 14:10:14.470432 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.565422 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rg7xl"] Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.573681 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.575924 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-4rn9v" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.583847 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.584016 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.603144 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rg7xl"] Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.611489 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9a0101-72b3-4879-a93b-f58362d175ed-combined-ca-bundle\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.611535 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-run-ovn\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.611564 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef9a0101-72b3-4879-a93b-f58362d175ed-scripts\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.611747 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef9a0101-72b3-4879-a93b-f58362d175ed-ovn-controller-tls-certs\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.611833 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-run\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.611992 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzp54\" (UniqueName: \"kubernetes.io/projected/ef9a0101-72b3-4879-a93b-f58362d175ed-kube-api-access-dzp54\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.612045 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-log-ovn\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.638254 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-cpcmr"] Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.640163 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.650115 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-cpcmr"] Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714467 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4jlf\" (UniqueName: \"kubernetes.io/projected/0ef872ad-6863-45f9-ac22-86841dea2bbf-kube-api-access-b4jlf\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714541 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9a0101-72b3-4879-a93b-f58362d175ed-combined-ca-bundle\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714577 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-run-ovn\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714618 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef9a0101-72b3-4879-a93b-f58362d175ed-scripts\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714648 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-log\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714698 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef9a0101-72b3-4879-a93b-f58362d175ed-ovn-controller-tls-certs\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714740 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-etc-ovs\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714776 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-run\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714859 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ef872ad-6863-45f9-ac22-86841dea2bbf-scripts\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714884 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzp54\" (UniqueName: \"kubernetes.io/projected/ef9a0101-72b3-4879-a93b-f58362d175ed-kube-api-access-dzp54\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714904 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-log-ovn\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714943 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-run\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.714970 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-lib\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.716440 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-run-ovn\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.717215 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-run\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.718525 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ef9a0101-72b3-4879-a93b-f58362d175ed-var-log-ovn\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.719200 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef9a0101-72b3-4879-a93b-f58362d175ed-scripts\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.719230 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9a0101-72b3-4879-a93b-f58362d175ed-combined-ca-bundle\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.720930 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef9a0101-72b3-4879-a93b-f58362d175ed-ovn-controller-tls-certs\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.733308 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzp54\" (UniqueName: \"kubernetes.io/projected/ef9a0101-72b3-4879-a93b-f58362d175ed-kube-api-access-dzp54\") pod \"ovn-controller-rg7xl\" (UID: \"ef9a0101-72b3-4879-a93b-f58362d175ed\") " pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.816430 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-run\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.816471 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-lib\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.816495 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4jlf\" (UniqueName: \"kubernetes.io/projected/0ef872ad-6863-45f9-ac22-86841dea2bbf-kube-api-access-b4jlf\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.816546 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-log\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.816597 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-etc-ovs\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.816682 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ef872ad-6863-45f9-ac22-86841dea2bbf-scripts\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.817043 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-run\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.817162 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-log\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.817197 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-etc-ovs\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.817291 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0ef872ad-6863-45f9-ac22-86841dea2bbf-var-lib\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.819321 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ef872ad-6863-45f9-ac22-86841dea2bbf-scripts\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.834612 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4jlf\" (UniqueName: \"kubernetes.io/projected/0ef872ad-6863-45f9-ac22-86841dea2bbf-kube-api-access-b4jlf\") pod \"ovn-controller-ovs-cpcmr\" (UID: \"0ef872ad-6863-45f9-ac22-86841dea2bbf\") " pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.917889 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rg7xl" Dec 04 14:10:15 crc kubenswrapper[4848]: I1204 14:10:15.967202 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:17 crc kubenswrapper[4848]: I1204 14:10:17.962022 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 14:10:17 crc kubenswrapper[4848]: I1204 14:10:17.964351 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:17 crc kubenswrapper[4848]: I1204 14:10:17.970359 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 04 14:10:17 crc kubenswrapper[4848]: I1204 14:10:17.970527 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 04 14:10:17 crc kubenswrapper[4848]: I1204 14:10:17.970693 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 04 14:10:17 crc kubenswrapper[4848]: I1204 14:10:17.970755 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tskvj" Dec 04 14:10:17 crc kubenswrapper[4848]: I1204 14:10:17.971668 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 04 14:10:17 crc kubenswrapper[4848]: I1204 14:10:17.974818 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.074223 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6mlj\" (UniqueName: \"kubernetes.io/projected/12b7748e-3169-49a0-a634-8a485f7690f4-kube-api-access-h6mlj\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.074298 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12b7748e-3169-49a0-a634-8a485f7690f4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.074322 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.074348 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.074367 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.074490 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/12b7748e-3169-49a0-a634-8a485f7690f4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.074530 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b7748e-3169-49a0-a634-8a485f7690f4-config\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.074733 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.176566 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.176990 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6mlj\" (UniqueName: \"kubernetes.io/projected/12b7748e-3169-49a0-a634-8a485f7690f4-kube-api-access-h6mlj\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.177023 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12b7748e-3169-49a0-a634-8a485f7690f4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.177040 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.177073 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.177093 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.177126 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/12b7748e-3169-49a0-a634-8a485f7690f4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.177167 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b7748e-3169-49a0-a634-8a485f7690f4-config\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.177512 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.178123 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12b7748e-3169-49a0-a634-8a485f7690f4-config\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.178276 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/12b7748e-3169-49a0-a634-8a485f7690f4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.178722 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12b7748e-3169-49a0-a634-8a485f7690f4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.182548 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.182889 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.183687 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b7748e-3169-49a0-a634-8a485f7690f4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.193831 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6mlj\" (UniqueName: \"kubernetes.io/projected/12b7748e-3169-49a0-a634-8a485f7690f4-kube-api-access-h6mlj\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.209273 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"12b7748e-3169-49a0-a634-8a485f7690f4\") " pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:18 crc kubenswrapper[4848]: I1204 14:10:18.286404 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.063461 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.065457 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.068230 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-tgjmn" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.070390 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.070581 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.070706 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.114368 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97bf1fd2-3a36-418d-826a-6dc60889b70c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.114442 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97bf1fd2-3a36-418d-826a-6dc60889b70c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.114489 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcjvr\" (UniqueName: \"kubernetes.io/projected/97bf1fd2-3a36-418d-826a-6dc60889b70c-kube-api-access-mcjvr\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.114527 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.114585 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.115443 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97bf1fd2-3a36-418d-826a-6dc60889b70c-config\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.115657 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.115774 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.125306 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217234 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97bf1fd2-3a36-418d-826a-6dc60889b70c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217286 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcjvr\" (UniqueName: \"kubernetes.io/projected/97bf1fd2-3a36-418d-826a-6dc60889b70c-kube-api-access-mcjvr\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217313 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217335 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217394 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97bf1fd2-3a36-418d-826a-6dc60889b70c-config\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217430 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217459 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217517 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97bf1fd2-3a36-418d-826a-6dc60889b70c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217941 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.217991 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97bf1fd2-3a36-418d-826a-6dc60889b70c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.218417 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97bf1fd2-3a36-418d-826a-6dc60889b70c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.219203 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97bf1fd2-3a36-418d-826a-6dc60889b70c-config\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.223769 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.223927 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.232272 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/97bf1fd2-3a36-418d-826a-6dc60889b70c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.235831 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcjvr\" (UniqueName: \"kubernetes.io/projected/97bf1fd2-3a36-418d-826a-6dc60889b70c-kube-api-access-mcjvr\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.250898 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"97bf1fd2-3a36-418d-826a-6dc60889b70c\") " pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:20 crc kubenswrapper[4848]: I1204 14:10:20.422377 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 14:10:21 crc kubenswrapper[4848]: I1204 14:10:21.690388 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 14:10:22 crc kubenswrapper[4848]: E1204 14:10:22.131548 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 14:10:22 crc kubenswrapper[4848]: E1204 14:10:22.131727 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m6lx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-nwfg8_openstack(02864bf6-0bad-49c3-a676-7117748f3ef6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:10:22 crc kubenswrapper[4848]: E1204 14:10:22.132872 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" podUID="02864bf6-0bad-49c3-a676-7117748f3ef6" Dec 04 14:10:22 crc kubenswrapper[4848]: E1204 14:10:22.170831 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 14:10:22 crc kubenswrapper[4848]: E1204 14:10:22.171093 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sq7t8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-znvrh_openstack(a2cf557a-089e-4c9d-b194-9ed01df93b6c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:10:22 crc kubenswrapper[4848]: E1204 14:10:22.172290 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" podUID="a2cf557a-089e-4c9d-b194-9ed01df93b6c" Dec 04 14:10:22 crc kubenswrapper[4848]: I1204 14:10:22.550771 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 14:10:23 crc kubenswrapper[4848]: W1204 14:10:23.286658 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadd25985_bc0c_48ee_9f42_4404503638b5.slice/crio-f7e12e2fe147faeefa7a63bf84e5a8b01f464042b78e94c302b5d76781184d9a WatchSource:0}: Error finding container f7e12e2fe147faeefa7a63bf84e5a8b01f464042b78e94c302b5d76781184d9a: Status 404 returned error can't find the container with id f7e12e2fe147faeefa7a63bf84e5a8b01f464042b78e94c302b5d76781184d9a Dec 04 14:10:23 crc kubenswrapper[4848]: W1204 14:10:23.333798 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4e5ee36_ea81_42cf_b0c4_44f2dcd4cd52.slice/crio-df174eda5dc161c1564646ca8a9b2f2191153f67c4cd32450fe1cad0b726fb7c WatchSource:0}: Error finding container df174eda5dc161c1564646ca8a9b2f2191153f67c4cd32450fe1cad0b726fb7c: Status 404 returned error can't find the container with id df174eda5dc161c1564646ca8a9b2f2191153f67c4cd32450fe1cad0b726fb7c Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.580860 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.624573 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.699724 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-dns-svc\") pod \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.700209 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq7t8\" (UniqueName: \"kubernetes.io/projected/a2cf557a-089e-4c9d-b194-9ed01df93b6c-kube-api-access-sq7t8\") pod \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.700376 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-config\") pod \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\" (UID: \"a2cf557a-089e-4c9d-b194-9ed01df93b6c\") " Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.701002 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2cf557a-089e-4c9d-b194-9ed01df93b6c" (UID: "a2cf557a-089e-4c9d-b194-9ed01df93b6c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.701067 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-config" (OuterVolumeSpecName: "config") pod "a2cf557a-089e-4c9d-b194-9ed01df93b6c" (UID: "a2cf557a-089e-4c9d-b194-9ed01df93b6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.709320 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2cf557a-089e-4c9d-b194-9ed01df93b6c-kube-api-access-sq7t8" (OuterVolumeSpecName: "kube-api-access-sq7t8") pod "a2cf557a-089e-4c9d-b194-9ed01df93b6c" (UID: "a2cf557a-089e-4c9d-b194-9ed01df93b6c"). InnerVolumeSpecName "kube-api-access-sq7t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.802873 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6lx6\" (UniqueName: \"kubernetes.io/projected/02864bf6-0bad-49c3-a676-7117748f3ef6-kube-api-access-m6lx6\") pod \"02864bf6-0bad-49c3-a676-7117748f3ef6\" (UID: \"02864bf6-0bad-49c3-a676-7117748f3ef6\") " Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.802923 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02864bf6-0bad-49c3-a676-7117748f3ef6-config\") pod \"02864bf6-0bad-49c3-a676-7117748f3ef6\" (UID: \"02864bf6-0bad-49c3-a676-7117748f3ef6\") " Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.804043 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.804063 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2cf557a-089e-4c9d-b194-9ed01df93b6c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.804075 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq7t8\" (UniqueName: \"kubernetes.io/projected/a2cf557a-089e-4c9d-b194-9ed01df93b6c-kube-api-access-sq7t8\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.804845 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02864bf6-0bad-49c3-a676-7117748f3ef6-config" (OuterVolumeSpecName: "config") pod "02864bf6-0bad-49c3-a676-7117748f3ef6" (UID: "02864bf6-0bad-49c3-a676-7117748f3ef6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.810525 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02864bf6-0bad-49c3-a676-7117748f3ef6-kube-api-access-m6lx6" (OuterVolumeSpecName: "kube-api-access-m6lx6") pod "02864bf6-0bad-49c3-a676-7117748f3ef6" (UID: "02864bf6-0bad-49c3-a676-7117748f3ef6"). InnerVolumeSpecName "kube-api-access-m6lx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.875326 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"add25985-bc0c-48ee-9f42-4404503638b5","Type":"ContainerStarted","Data":"f7e12e2fe147faeefa7a63bf84e5a8b01f464042b78e94c302b5d76781184d9a"} Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.878027 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52","Type":"ContainerStarted","Data":"df174eda5dc161c1564646ca8a9b2f2191153f67c4cd32450fe1cad0b726fb7c"} Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.879313 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" event={"ID":"a2cf557a-089e-4c9d-b194-9ed01df93b6c","Type":"ContainerDied","Data":"c22dc559b743483e65872621de4b84d2a90a889500ab51128fd456342afe7ecd"} Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.879393 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-znvrh" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.883257 4848 generic.go:334] "Generic (PLEG): container finished" podID="f66fc099-8a2e-49ea-803c-417aeedbe389" containerID="8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc" exitCode=0 Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.883295 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" event={"ID":"f66fc099-8a2e-49ea-803c-417aeedbe389","Type":"ContainerDied","Data":"8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc"} Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.886356 4848 generic.go:334] "Generic (PLEG): container finished" podID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerID="37c4c43e23ca2eb4d202c76e4e5d67827148d0ca89db849fe53ec2c92e1ff5f8" exitCode=0 Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.886413 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" event={"ID":"c1c86a95-c98a-4049-901d-c2ac744d42f8","Type":"ContainerDied","Data":"37c4c43e23ca2eb4d202c76e4e5d67827148d0ca89db849fe53ec2c92e1ff5f8"} Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.888341 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" event={"ID":"02864bf6-0bad-49c3-a676-7117748f3ef6","Type":"ContainerDied","Data":"f63a7aba2da8fd52deb26bf84a96bcb2db8ce0b6926f95c5a2b30ea60b710a9c"} Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.888455 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-nwfg8" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.909261 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6lx6\" (UniqueName: \"kubernetes.io/projected/02864bf6-0bad-49c3-a676-7117748f3ef6-kube-api-access-m6lx6\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:23 crc kubenswrapper[4848]: I1204 14:10:23.909300 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02864bf6-0bad-49c3-a676-7117748f3ef6-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.011670 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwfg8"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.023459 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-nwfg8"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.042522 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-znvrh"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.059099 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-znvrh"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.065364 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.432995 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02864bf6-0bad-49c3-a676-7117748f3ef6" path="/var/lib/kubelet/pods/02864bf6-0bad-49c3-a676-7117748f3ef6/volumes" Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.433459 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2cf557a-089e-4c9d-b194-9ed01df93b6c" path="/var/lib/kubelet/pods/a2cf557a-089e-4c9d-b194-9ed01df93b6c/volumes" Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.560596 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.572669 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:10:24 crc kubenswrapper[4848]: W1204 14:10:24.594849 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8efe0158_56d4_4526_9683_263fb1ce6f23.slice/crio-3e245660c98c03acdf68abdb74fdc25a9fc63f422b0059bd5147afe889fcc7e3 WatchSource:0}: Error finding container 3e245660c98c03acdf68abdb74fdc25a9fc63f422b0059bd5147afe889fcc7e3: Status 404 returned error can't find the container with id 3e245660c98c03acdf68abdb74fdc25a9fc63f422b0059bd5147afe889fcc7e3 Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.768784 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.779306 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.791892 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5db675fff6-2jrj6"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.801116 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rg7xl"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.884644 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.900223 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerStarted","Data":"d2a9bf9f9b8cd012e4d8789f8c5b5f0f0eb339ecdb3c505e82c537a4727ec528"} Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.901417 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6049ed43-40db-4bff-b89b-88cce285fd2d","Type":"ContainerStarted","Data":"b2e2421f5614069784bc785cfd645ef5c68ba2126d050bdd27b12f15716d4f92"} Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.902222 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" event={"ID":"8efe0158-56d4-4526-9683-263fb1ce6f23","Type":"ContainerStarted","Data":"3e245660c98c03acdf68abdb74fdc25a9fc63f422b0059bd5147afe889fcc7e3"} Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.903790 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" event={"ID":"f66fc099-8a2e-49ea-803c-417aeedbe389","Type":"ContainerStarted","Data":"c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012"} Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.906310 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" event={"ID":"c1c86a95-c98a-4049-901d-c2ac744d42f8","Type":"ContainerStarted","Data":"3912b870fc3fd6cfc6730d6676bc8289b11140b35da436e896471f3b6b5d6302"} Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.906665 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.930608 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" podStartSLOduration=2.5969341310000003 podStartE2EDuration="18.930593549s" podCreationTimestamp="2025-12-04 14:10:06 +0000 UTC" firstStartedPulling="2025-12-04 14:10:07.138472963 +0000 UTC m=+1311.080969481" lastFinishedPulling="2025-12-04 14:10:23.472132371 +0000 UTC m=+1327.414628899" observedRunningTime="2025-12-04 14:10:24.929397231 +0000 UTC m=+1328.871893769" watchObservedRunningTime="2025-12-04 14:10:24.930593549 +0000 UTC m=+1328.873090077" Dec 04 14:10:24 crc kubenswrapper[4848]: I1204 14:10:24.954212 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" podStartSLOduration=2.827445315 podStartE2EDuration="18.954187391s" podCreationTimestamp="2025-12-04 14:10:06 +0000 UTC" firstStartedPulling="2025-12-04 14:10:07.332248336 +0000 UTC m=+1311.274744864" lastFinishedPulling="2025-12-04 14:10:23.458990412 +0000 UTC m=+1327.401486940" observedRunningTime="2025-12-04 14:10:24.949714623 +0000 UTC m=+1328.892211171" watchObservedRunningTime="2025-12-04 14:10:24.954187391 +0000 UTC m=+1328.896683929" Dec 04 14:10:25 crc kubenswrapper[4848]: I1204 14:10:25.512045 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 14:10:25 crc kubenswrapper[4848]: I1204 14:10:25.663575 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-cpcmr"] Dec 04 14:10:25 crc kubenswrapper[4848]: I1204 14:10:25.925326 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30883c46-53ec-45f9-84f6-29a64ccf8994","Type":"ContainerStarted","Data":"0f318664aee27c7d28309de918f73ef54729735695edff04d26277043b9fe4de"} Dec 04 14:10:25 crc kubenswrapper[4848]: I1204 14:10:25.929817 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6049ed43-40db-4bff-b89b-88cce285fd2d","Type":"ContainerStarted","Data":"c4829738b3394f57da24a7580486aa1ca47db4fc2c06fe141440e1c598436e14"} Dec 04 14:10:25 crc kubenswrapper[4848]: I1204 14:10:25.930268 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:26 crc kubenswrapper[4848]: I1204 14:10:26.941771 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5db675fff6-2jrj6" event={"ID":"6167e58f-c3af-424d-9687-0c5266be601f","Type":"ContainerStarted","Data":"b81e4c8284e3b02b25d95a9e2d6ae96894cf1fae1ec4e770ba261ceaf99eb50a"} Dec 04 14:10:26 crc kubenswrapper[4848]: I1204 14:10:26.944562 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2","Type":"ContainerStarted","Data":"a24cf3d7a4f9a1daaeab94d50529164cccc8987e2a1120c91dc7bf173caa6748"} Dec 04 14:10:26 crc kubenswrapper[4848]: I1204 14:10:26.945918 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"12b7748e-3169-49a0-a634-8a485f7690f4","Type":"ContainerStarted","Data":"a9e149f204fe3027d100d4763ac580fb62b41b45206ab97b87bc2b82f1a8487f"} Dec 04 14:10:26 crc kubenswrapper[4848]: I1204 14:10:26.947179 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rg7xl" event={"ID":"ef9a0101-72b3-4879-a93b-f58362d175ed","Type":"ContainerStarted","Data":"437dc4923afb860d9b22d219ea725df9026658dfb3f86dcb1df8831a53b5bae3"} Dec 04 14:10:26 crc kubenswrapper[4848]: I1204 14:10:26.948482 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d644fdb6-4d43-4e46-9537-f2fedef790e0","Type":"ContainerStarted","Data":"f0933b6194a199f8f21e481395def7a57d5dd68b9b1c03c32ed20d2a1b50b399"} Dec 04 14:10:30 crc kubenswrapper[4848]: W1204 14:10:30.508860 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97bf1fd2_3a36_418d_826a_6dc60889b70c.slice/crio-a056d5a2a19a3ea51f4b6515a2eb28571780ddd5f00a61b9628a21a1056794df WatchSource:0}: Error finding container a056d5a2a19a3ea51f4b6515a2eb28571780ddd5f00a61b9628a21a1056794df: Status 404 returned error can't find the container with id a056d5a2a19a3ea51f4b6515a2eb28571780ddd5f00a61b9628a21a1056794df Dec 04 14:10:30 crc kubenswrapper[4848]: W1204 14:10:30.516093 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ef872ad_6863_45f9_ac22_86841dea2bbf.slice/crio-abf88921a91fa156e0291f095f88c7251adb4776b3d7b4f66fdc8fdd54b9efee WatchSource:0}: Error finding container abf88921a91fa156e0291f095f88c7251adb4776b3d7b4f66fdc8fdd54b9efee: Status 404 returned error can't find the container with id abf88921a91fa156e0291f095f88c7251adb4776b3d7b4f66fdc8fdd54b9efee Dec 04 14:10:30 crc kubenswrapper[4848]: I1204 14:10:30.990111 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"97bf1fd2-3a36-418d-826a-6dc60889b70c","Type":"ContainerStarted","Data":"a056d5a2a19a3ea51f4b6515a2eb28571780ddd5f00a61b9628a21a1056794df"} Dec 04 14:10:30 crc kubenswrapper[4848]: I1204 14:10:30.991532 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cpcmr" event={"ID":"0ef872ad-6863-45f9-ac22-86841dea2bbf","Type":"ContainerStarted","Data":"abf88921a91fa156e0291f095f88c7251adb4776b3d7b4f66fdc8fdd54b9efee"} Dec 04 14:10:31 crc kubenswrapper[4848]: I1204 14:10:31.441796 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:31 crc kubenswrapper[4848]: I1204 14:10:31.833111 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:31 crc kubenswrapper[4848]: I1204 14:10:31.918301 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-grcrh"] Dec 04 14:10:32 crc kubenswrapper[4848]: I1204 14:10:32.004059 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerName="dnsmasq-dns" containerID="cri-o://3912b870fc3fd6cfc6730d6676bc8289b11140b35da436e896471f3b6b5d6302" gracePeriod=10 Dec 04 14:10:32 crc kubenswrapper[4848]: I1204 14:10:32.005160 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5db675fff6-2jrj6" event={"ID":"6167e58f-c3af-424d-9687-0c5266be601f","Type":"ContainerStarted","Data":"705ca816f506b8e035a94a4f05e15f2ddb552353361ff9c434fbd9c5c8abd819"} Dec 04 14:10:32 crc kubenswrapper[4848]: I1204 14:10:32.065345 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5db675fff6-2jrj6" podStartSLOduration=19.065322137 podStartE2EDuration="19.065322137s" podCreationTimestamp="2025-12-04 14:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:10:32.057397545 +0000 UTC m=+1335.999894073" watchObservedRunningTime="2025-12-04 14:10:32.065322137 +0000 UTC m=+1336.007818655" Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.014305 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2","Type":"ContainerStarted","Data":"ebeba60f20961bcb98e21f3518ae09ccda10ac7c18ef69412e1cad35566b43ea"} Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.016440 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" event={"ID":"8efe0158-56d4-4526-9683-263fb1ce6f23","Type":"ContainerStarted","Data":"4a09e7f53c8e73e2c0bb63f5368ad4c65d7d885f1097e48df423e39d6a028611"} Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.018794 4848 generic.go:334] "Generic (PLEG): container finished" podID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerID="3912b870fc3fd6cfc6730d6676bc8289b11140b35da436e896471f3b6b5d6302" exitCode=0 Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.018854 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" event={"ID":"c1c86a95-c98a-4049-901d-c2ac744d42f8","Type":"ContainerDied","Data":"3912b870fc3fd6cfc6730d6676bc8289b11140b35da436e896471f3b6b5d6302"} Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.020333 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"add25985-bc0c-48ee-9f42-4404503638b5","Type":"ContainerStarted","Data":"a473dd5892dc15fdd133c9bca42fb64ec74226d9ed0494f3ee5a340903d70697"} Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.020675 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.022626 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52","Type":"ContainerStarted","Data":"fd05c94679406b35d3d1c06b385591e89ccaf9d97dbdd825357e1f67027e525f"} Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.072155 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=15.0267562 podStartE2EDuration="23.072137556s" podCreationTimestamp="2025-12-04 14:10:10 +0000 UTC" firstStartedPulling="2025-12-04 14:10:23.31108214 +0000 UTC m=+1327.253578668" lastFinishedPulling="2025-12-04 14:10:31.356463496 +0000 UTC m=+1335.298960024" observedRunningTime="2025-12-04 14:10:33.061414486 +0000 UTC m=+1337.003911014" watchObservedRunningTime="2025-12-04 14:10:33.072137556 +0000 UTC m=+1337.014634084" Dec 04 14:10:33 crc kubenswrapper[4848]: I1204 14:10:33.100979 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-49kl7" podStartSLOduration=13.182101505 podStartE2EDuration="20.100960784s" podCreationTimestamp="2025-12-04 14:10:13 +0000 UTC" firstStartedPulling="2025-12-04 14:10:24.599675114 +0000 UTC m=+1328.542171632" lastFinishedPulling="2025-12-04 14:10:31.518534383 +0000 UTC m=+1335.461030911" observedRunningTime="2025-12-04 14:10:33.091835533 +0000 UTC m=+1337.034332051" watchObservedRunningTime="2025-12-04 14:10:33.100960784 +0000 UTC m=+1337.043457312" Dec 04 14:10:34 crc kubenswrapper[4848]: I1204 14:10:34.048795 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:34 crc kubenswrapper[4848]: I1204 14:10:34.049212 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:34 crc kubenswrapper[4848]: I1204 14:10:34.054330 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:35 crc kubenswrapper[4848]: I1204 14:10:35.050901 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5db675fff6-2jrj6" Dec 04 14:10:35 crc kubenswrapper[4848]: I1204 14:10:35.136620 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-8554695464-z4rxq"] Dec 04 14:10:36 crc kubenswrapper[4848]: I1204 14:10:36.467070 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Dec 04 14:10:39 crc kubenswrapper[4848]: I1204 14:10:39.089106 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerStarted","Data":"82a8f10da12cd1dca25f294fd60d46653a7a4b4a0c6c4c7c22457cd12570dea7"} Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.095112 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.757584 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.891248 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmszc\" (UniqueName: \"kubernetes.io/projected/c1c86a95-c98a-4049-901d-c2ac744d42f8-kube-api-access-nmszc\") pod \"c1c86a95-c98a-4049-901d-c2ac744d42f8\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.891436 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-config\") pod \"c1c86a95-c98a-4049-901d-c2ac744d42f8\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.891545 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-dns-svc\") pod \"c1c86a95-c98a-4049-901d-c2ac744d42f8\" (UID: \"c1c86a95-c98a-4049-901d-c2ac744d42f8\") " Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.898184 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1c86a95-c98a-4049-901d-c2ac744d42f8-kube-api-access-nmszc" (OuterVolumeSpecName: "kube-api-access-nmszc") pod "c1c86a95-c98a-4049-901d-c2ac744d42f8" (UID: "c1c86a95-c98a-4049-901d-c2ac744d42f8"). InnerVolumeSpecName "kube-api-access-nmszc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.937849 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-config" (OuterVolumeSpecName: "config") pod "c1c86a95-c98a-4049-901d-c2ac744d42f8" (UID: "c1c86a95-c98a-4049-901d-c2ac744d42f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.938267 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c1c86a95-c98a-4049-901d-c2ac744d42f8" (UID: "c1c86a95-c98a-4049-901d-c2ac744d42f8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.993599 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmszc\" (UniqueName: \"kubernetes.io/projected/c1c86a95-c98a-4049-901d-c2ac744d42f8-kube-api-access-nmszc\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.993864 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:41 crc kubenswrapper[4848]: I1204 14:10:41.993927 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c86a95-c98a-4049-901d-c2ac744d42f8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:42 crc kubenswrapper[4848]: I1204 14:10:42.120549 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" event={"ID":"c1c86a95-c98a-4049-901d-c2ac744d42f8","Type":"ContainerDied","Data":"5dc87611d6e63349296355dfa51b2e14b790d6e5d1b8c5513f334288d58e5f73"} Dec 04 14:10:42 crc kubenswrapper[4848]: I1204 14:10:42.120921 4848 scope.go:117] "RemoveContainer" containerID="3912b870fc3fd6cfc6730d6676bc8289b11140b35da436e896471f3b6b5d6302" Dec 04 14:10:42 crc kubenswrapper[4848]: I1204 14:10:42.121054 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" Dec 04 14:10:42 crc kubenswrapper[4848]: I1204 14:10:42.160164 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-grcrh"] Dec 04 14:10:42 crc kubenswrapper[4848]: I1204 14:10:42.168896 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-grcrh"] Dec 04 14:10:42 crc kubenswrapper[4848]: E1204 14:10:42.251786 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Dec 04 14:10:42 crc kubenswrapper[4848]: E1204 14:10:42.252031 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n55hcfh548h7fhdch564hf6hb6h86h5dh59dh9bh685h5b8h5dh64chd5h5cch6bh5d8h5b8h648h649hbbh69h57bh5f6hfch5ddh68bh5b7h548q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h6mlj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(12b7748e-3169-49a0-a634-8a485f7690f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:10:42 crc kubenswrapper[4848]: I1204 14:10:42.407893 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" path="/var/lib/kubelet/pods/c1c86a95-c98a-4049-901d-c2ac744d42f8/volumes" Dec 04 14:10:42 crc kubenswrapper[4848]: E1204 14:10:42.476397 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Dec 04 14:10:42 crc kubenswrapper[4848]: E1204 14:10:42.476735 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n674h68fh544h97h5f7h9h7fh547h689h59bh5ddh65ch5ffh655hdbh7dh79h88h5b4h7dh68dh656hcdh56dh575h556h588h696h54dh586h565h569q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dzp54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-rg7xl_openstack(ef9a0101-72b3-4879-a93b-f58362d175ed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:10:42 crc kubenswrapper[4848]: E1204 14:10:42.478366 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-rg7xl" podUID="ef9a0101-72b3-4879-a93b-f58362d175ed" Dec 04 14:10:42 crc kubenswrapper[4848]: E1204 14:10:42.653749 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Dec 04 14:10:42 crc kubenswrapper[4848]: E1204 14:10:42.654087 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n596h57dh8dh55bh5b8hc7hc4h58bh695h59ch5fch694h548h68bh564h5cch86h65fh5ddh57bhf8h5f6h78h58fhf9h5c5hc7h68bh5cch8fhcbh5b6q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mcjvr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(97bf1fd2-3a36-418d-826a-6dc60889b70c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:10:42 crc kubenswrapper[4848]: I1204 14:10:42.889038 4848 scope.go:117] "RemoveContainer" containerID="37c4c43e23ca2eb4d202c76e4e5d67827148d0ca89db849fe53ec2c92e1ff5f8" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.080874 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-2hkgw"] Dec 04 14:10:43 crc kubenswrapper[4848]: E1204 14:10:43.081421 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerName="dnsmasq-dns" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.081448 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerName="dnsmasq-dns" Dec 04 14:10:43 crc kubenswrapper[4848]: E1204 14:10:43.081471 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerName="init" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.081482 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerName="init" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.086598 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerName="dnsmasq-dns" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.088137 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.092087 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-2hkgw"] Dec 04 14:10:43 crc kubenswrapper[4848]: E1204 14:10:43.136253 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-rg7xl" podUID="ef9a0101-72b3-4879-a93b-f58362d175ed" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.216625 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.216744 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk7w4\" (UniqueName: \"kubernetes.io/projected/565e9c18-f099-49a4-9457-725b85e0cb7a-kube-api-access-vk7w4\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.216791 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-config\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.318347 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk7w4\" (UniqueName: \"kubernetes.io/projected/565e9c18-f099-49a4-9457-725b85e0cb7a-kube-api-access-vk7w4\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.318444 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-config\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.318566 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.319388 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-config\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.319918 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.339477 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk7w4\" (UniqueName: \"kubernetes.io/projected/565e9c18-f099-49a4-9457-725b85e0cb7a-kube-api-access-vk7w4\") pod \"dnsmasq-dns-7cb5889db5-2hkgw\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:43 crc kubenswrapper[4848]: I1204 14:10:43.410149 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.201934 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.215511 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.217989 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.218063 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.218134 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.218242 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-tw6tt" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.226084 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.314718 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.314775 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.353195 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrj5k\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-kube-api-access-rrj5k\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.353236 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.353296 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.353384 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0102871d-c950-46e7-af43-a2b3f1cbc39f-lock\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.353430 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0102871d-c950-46e7-af43-a2b3f1cbc39f-cache\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.455239 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.455327 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0102871d-c950-46e7-af43-a2b3f1cbc39f-lock\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.455427 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0102871d-c950-46e7-af43-a2b3f1cbc39f-cache\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.455512 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrj5k\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-kube-api-access-rrj5k\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.455550 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: E1204 14:10:44.455815 4848 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:10:44 crc kubenswrapper[4848]: E1204 14:10:44.455897 4848 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:10:44 crc kubenswrapper[4848]: E1204 14:10:44.455943 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift podName:0102871d-c950-46e7-af43-a2b3f1cbc39f nodeName:}" failed. No retries permitted until 2025-12-04 14:10:44.955927531 +0000 UTC m=+1348.898424059 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift") pod "swift-storage-0" (UID: "0102871d-c950-46e7-af43-a2b3f1cbc39f") : configmap "swift-ring-files" not found Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.456075 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.456691 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0102871d-c950-46e7-af43-a2b3f1cbc39f-lock\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.456787 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0102871d-c950-46e7-af43-a2b3f1cbc39f-cache\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.476338 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrj5k\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-kube-api-access-rrj5k\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.480083 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.715726 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-2hkgw"] Dec 04 14:10:44 crc kubenswrapper[4848]: I1204 14:10:44.964719 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:44 crc kubenswrapper[4848]: E1204 14:10:44.964926 4848 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:10:44 crc kubenswrapper[4848]: E1204 14:10:44.964967 4848 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:10:44 crc kubenswrapper[4848]: E1204 14:10:44.965022 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift podName:0102871d-c950-46e7-af43-a2b3f1cbc39f nodeName:}" failed. No retries permitted until 2025-12-04 14:10:45.965004502 +0000 UTC m=+1349.907501020 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift") pod "swift-storage-0" (UID: "0102871d-c950-46e7-af43-a2b3f1cbc39f") : configmap "swift-ring-files" not found Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.157326 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cpcmr" event={"ID":"0ef872ad-6863-45f9-ac22-86841dea2bbf","Type":"ContainerStarted","Data":"8e0b5e910d2577862317eee6e8f3d4ff8a26096d39b2a1161277c2f7cc6222d4"} Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.162706 4848 generic.go:334] "Generic (PLEG): container finished" podID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerID="82a8f10da12cd1dca25f294fd60d46653a7a4b4a0c6c4c7c22457cd12570dea7" exitCode=0 Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.162770 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerDied","Data":"82a8f10da12cd1dca25f294fd60d46653a7a4b4a0c6c4c7c22457cd12570dea7"} Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.165352 4848 generic.go:334] "Generic (PLEG): container finished" podID="f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52" containerID="fd05c94679406b35d3d1c06b385591e89ccaf9d97dbdd825357e1f67027e525f" exitCode=0 Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.165409 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52","Type":"ContainerDied","Data":"fd05c94679406b35d3d1c06b385591e89ccaf9d97dbdd825357e1f67027e525f"} Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.169132 4848 generic.go:334] "Generic (PLEG): container finished" podID="1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2" containerID="ebeba60f20961bcb98e21f3518ae09ccda10ac7c18ef69412e1cad35566b43ea" exitCode=0 Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.169210 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2","Type":"ContainerDied","Data":"ebeba60f20961bcb98e21f3518ae09ccda10ac7c18ef69412e1cad35566b43ea"} Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.171425 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d644fdb6-4d43-4e46-9537-f2fedef790e0","Type":"ContainerStarted","Data":"b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d"} Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.171567 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.234347 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=14.896120712 podStartE2EDuration="33.234331916s" podCreationTimestamp="2025-12-04 14:10:12 +0000 UTC" firstStartedPulling="2025-12-04 14:10:26.187390164 +0000 UTC m=+1330.129886732" lastFinishedPulling="2025-12-04 14:10:44.525601408 +0000 UTC m=+1348.468097936" observedRunningTime="2025-12-04 14:10:45.22294195 +0000 UTC m=+1349.165438498" watchObservedRunningTime="2025-12-04 14:10:45.234331916 +0000 UTC m=+1349.176828444" Dec 04 14:10:45 crc kubenswrapper[4848]: I1204 14:10:45.990011 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:45 crc kubenswrapper[4848]: E1204 14:10:45.990572 4848 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:10:45 crc kubenswrapper[4848]: E1204 14:10:45.990687 4848 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:10:45 crc kubenswrapper[4848]: E1204 14:10:45.990857 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift podName:0102871d-c950-46e7-af43-a2b3f1cbc39f nodeName:}" failed. No retries permitted until 2025-12-04 14:10:47.990834901 +0000 UTC m=+1351.933331439 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift") pod "swift-storage-0" (UID: "0102871d-c950-46e7-af43-a2b3f1cbc39f") : configmap "swift-ring-files" not found Dec 04 14:10:46 crc kubenswrapper[4848]: I1204 14:10:46.182676 4848 generic.go:334] "Generic (PLEG): container finished" podID="0ef872ad-6863-45f9-ac22-86841dea2bbf" containerID="8e0b5e910d2577862317eee6e8f3d4ff8a26096d39b2a1161277c2f7cc6222d4" exitCode=0 Dec 04 14:10:46 crc kubenswrapper[4848]: I1204 14:10:46.182817 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cpcmr" event={"ID":"0ef872ad-6863-45f9-ac22-86841dea2bbf","Type":"ContainerDied","Data":"8e0b5e910d2577862317eee6e8f3d4ff8a26096d39b2a1161277c2f7cc6222d4"} Dec 04 14:10:46 crc kubenswrapper[4848]: I1204 14:10:46.443828 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5ccc8479f9-grcrh" podUID="c1c86a95-c98a-4049-901d-c2ac744d42f8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Dec 04 14:10:46 crc kubenswrapper[4848]: W1204 14:10:46.696617 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod565e9c18_f099_49a4_9457_725b85e0cb7a.slice/crio-f5d8714e32ed12c5b1638fef3415452f8a9e50add47bdd61a4ebe0af7b09c532 WatchSource:0}: Error finding container f5d8714e32ed12c5b1638fef3415452f8a9e50add47bdd61a4ebe0af7b09c532: Status 404 returned error can't find the container with id f5d8714e32ed12c5b1638fef3415452f8a9e50add47bdd61a4ebe0af7b09c532 Dec 04 14:10:47 crc kubenswrapper[4848]: E1204 14:10:47.102067 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="97bf1fd2-3a36-418d-826a-6dc60889b70c" Dec 04 14:10:47 crc kubenswrapper[4848]: E1204 14:10:47.110990 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="12b7748e-3169-49a0-a634-8a485f7690f4" Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.208791 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"12b7748e-3169-49a0-a634-8a485f7690f4","Type":"ContainerStarted","Data":"d58bec5a31af83833792e854258204521e8c227a66287ebdb70bced072a984d1"} Dec 04 14:10:47 crc kubenswrapper[4848]: E1204 14:10:47.211526 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="12b7748e-3169-49a0-a634-8a485f7690f4" Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.211701 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"97bf1fd2-3a36-418d-826a-6dc60889b70c","Type":"ContainerStarted","Data":"d70d9522c2d395cab8079d5c91602c7d3b5250ca5b109074406853c271774855"} Dec 04 14:10:47 crc kubenswrapper[4848]: E1204 14:10:47.213577 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="97bf1fd2-3a36-418d-826a-6dc60889b70c" Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.215774 4848 generic.go:334] "Generic (PLEG): container finished" podID="565e9c18-f099-49a4-9457-725b85e0cb7a" containerID="66e54b150809137932b9d4b609fafb607f0bc4d29ae529f038da48b097ecbd3c" exitCode=0 Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.215867 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" event={"ID":"565e9c18-f099-49a4-9457-725b85e0cb7a","Type":"ContainerDied","Data":"66e54b150809137932b9d4b609fafb607f0bc4d29ae529f038da48b097ecbd3c"} Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.215891 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" event={"ID":"565e9c18-f099-49a4-9457-725b85e0cb7a","Type":"ContainerStarted","Data":"f5d8714e32ed12c5b1638fef3415452f8a9e50add47bdd61a4ebe0af7b09c532"} Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.218696 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cpcmr" event={"ID":"0ef872ad-6863-45f9-ac22-86841dea2bbf","Type":"ContainerStarted","Data":"946736b799da2a8e70a212c2012203c66702145acdc401f41db923b330d35905"} Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.226175 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52","Type":"ContainerStarted","Data":"fa0d4cb7328dceab371b944208cefc639d795a83a40e968eea05a84e48cbb76d"} Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.238755 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2","Type":"ContainerStarted","Data":"8bc7162b6f03e03cc234a466c5264c321dc9dc924cd80822094847c769fc596e"} Dec 04 14:10:47 crc kubenswrapper[4848]: I1204 14:10:47.316573 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=31.388885399 podStartE2EDuration="39.316533244s" podCreationTimestamp="2025-12-04 14:10:08 +0000 UTC" firstStartedPulling="2025-12-04 14:10:23.359846901 +0000 UTC m=+1327.302343429" lastFinishedPulling="2025-12-04 14:10:31.287494746 +0000 UTC m=+1335.229991274" observedRunningTime="2025-12-04 14:10:47.291791344 +0000 UTC m=+1351.234287872" watchObservedRunningTime="2025-12-04 14:10:47.316533244 +0000 UTC m=+1351.259029772" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.041200 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:48 crc kubenswrapper[4848]: E1204 14:10:48.041451 4848 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:10:48 crc kubenswrapper[4848]: E1204 14:10:48.041602 4848 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:10:48 crc kubenswrapper[4848]: E1204 14:10:48.041679 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift podName:0102871d-c950-46e7-af43-a2b3f1cbc39f nodeName:}" failed. No retries permitted until 2025-12-04 14:10:52.041655649 +0000 UTC m=+1355.984152207 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift") pod "swift-storage-0" (UID: "0102871d-c950-46e7-af43-a2b3f1cbc39f") : configmap "swift-ring-files" not found Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.205435 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=34.092128384 podStartE2EDuration="39.205415406s" podCreationTimestamp="2025-12-04 14:10:09 +0000 UTC" firstStartedPulling="2025-12-04 14:10:26.19963374 +0000 UTC m=+1330.142130268" lastFinishedPulling="2025-12-04 14:10:31.312920762 +0000 UTC m=+1335.255417290" observedRunningTime="2025-12-04 14:10:47.332348347 +0000 UTC m=+1351.274844885" watchObservedRunningTime="2025-12-04 14:10:48.205415406 +0000 UTC m=+1352.147911934" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.256659 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9ltlr"] Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.258404 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.263904 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.265205 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.265549 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" event={"ID":"565e9c18-f099-49a4-9457-725b85e0cb7a","Type":"ContainerStarted","Data":"def06344ecd9560b993c752d7c61cedd7d29be1747532632fd15618959d6afaf"} Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.265616 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.270610 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cpcmr" event={"ID":"0ef872ad-6863-45f9-ac22-86841dea2bbf","Type":"ContainerStarted","Data":"1dcfa8ad7972adfd63fb4e471e9b87274e201befbcb4bddfaf97ac14bcf83779"} Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.270657 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.274596 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:10:48 crc kubenswrapper[4848]: E1204 14:10:48.274776 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="12b7748e-3169-49a0-a634-8a485f7690f4" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.274936 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 14:10:48 crc kubenswrapper[4848]: E1204 14:10:48.275359 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="97bf1fd2-3a36-418d-826a-6dc60889b70c" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.321513 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9ltlr"] Dec 04 14:10:48 crc kubenswrapper[4848]: E1204 14:10:48.322356 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-cd9lz ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-cd9lz ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-9ltlr" podUID="6e5c2028-7096-441e-a33d-60cc6f23880c" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.353434 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9ltlr"] Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.361992 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-ring-data-devices\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.362029 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd9lz\" (UniqueName: \"kubernetes.io/projected/6e5c2028-7096-441e-a33d-60cc6f23880c-kube-api-access-cd9lz\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.362054 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e5c2028-7096-441e-a33d-60cc6f23880c-etc-swift\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.362274 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-scripts\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.362419 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-dispersionconf\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.362573 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-swiftconf\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.362658 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-combined-ca-bundle\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.377394 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-gplgt"] Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.379151 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gplgt"] Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.379236 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.410842 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" podStartSLOduration=5.410821441 podStartE2EDuration="5.410821441s" podCreationTimestamp="2025-12-04 14:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:10:48.315483072 +0000 UTC m=+1352.257979600" watchObservedRunningTime="2025-12-04 14:10:48.410821441 +0000 UTC m=+1352.353317969" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.420933 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-cpcmr" podStartSLOduration=20.735218595 podStartE2EDuration="33.420911556s" podCreationTimestamp="2025-12-04 14:10:15 +0000 UTC" firstStartedPulling="2025-12-04 14:10:30.569941024 +0000 UTC m=+1334.512437552" lastFinishedPulling="2025-12-04 14:10:43.255633995 +0000 UTC m=+1347.198130513" observedRunningTime="2025-12-04 14:10:48.337364352 +0000 UTC m=+1352.279860880" watchObservedRunningTime="2025-12-04 14:10:48.420911556 +0000 UTC m=+1352.363408084" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.464445 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-combined-ca-bundle\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.464541 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ff826431-be61-4dea-82e7-b6849d838e52-etc-swift\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.464590 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-swiftconf\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.464648 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-combined-ca-bundle\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.464864 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-ring-data-devices\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.464892 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd9lz\" (UniqueName: \"kubernetes.io/projected/6e5c2028-7096-441e-a33d-60cc6f23880c-kube-api-access-cd9lz\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.464920 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-ring-data-devices\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.464966 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e5c2028-7096-441e-a33d-60cc6f23880c-etc-swift\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465060 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-dispersionconf\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465101 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-scripts\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465180 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-scripts\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465313 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-dispersionconf\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465373 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-swiftconf\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465419 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6zqx\" (UniqueName: \"kubernetes.io/projected/ff826431-be61-4dea-82e7-b6849d838e52-kube-api-access-f6zqx\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465777 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-ring-data-devices\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465780 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e5c2028-7096-441e-a33d-60cc6f23880c-etc-swift\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.465971 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-scripts\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.474464 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-swiftconf\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.474620 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-combined-ca-bundle\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.480336 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-dispersionconf\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.485990 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd9lz\" (UniqueName: \"kubernetes.io/projected/6e5c2028-7096-441e-a33d-60cc6f23880c-kube-api-access-cd9lz\") pod \"swift-ring-rebalance-9ltlr\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.567066 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-swiftconf\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.567905 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6zqx\" (UniqueName: \"kubernetes.io/projected/ff826431-be61-4dea-82e7-b6849d838e52-kube-api-access-f6zqx\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.568774 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-combined-ca-bundle\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.568824 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ff826431-be61-4dea-82e7-b6849d838e52-etc-swift\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.568878 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-ring-data-devices\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.568925 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-dispersionconf\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.568993 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-scripts\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.569529 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ff826431-be61-4dea-82e7-b6849d838e52-etc-swift\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.569944 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-scripts\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.570025 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-ring-data-devices\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.571741 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-swiftconf\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.572433 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-dispersionconf\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.572994 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-combined-ca-bundle\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.590934 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6zqx\" (UniqueName: \"kubernetes.io/projected/ff826431-be61-4dea-82e7-b6849d838e52-kube-api-access-f6zqx\") pod \"swift-ring-rebalance-gplgt\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:48 crc kubenswrapper[4848]: I1204 14:10:48.708807 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.279426 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.295094 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.383117 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-ring-data-devices\") pod \"6e5c2028-7096-441e-a33d-60cc6f23880c\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.383175 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-combined-ca-bundle\") pod \"6e5c2028-7096-441e-a33d-60cc6f23880c\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.383205 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-scripts\") pod \"6e5c2028-7096-441e-a33d-60cc6f23880c\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.383260 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e5c2028-7096-441e-a33d-60cc6f23880c-etc-swift\") pod \"6e5c2028-7096-441e-a33d-60cc6f23880c\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.383281 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-swiftconf\") pod \"6e5c2028-7096-441e-a33d-60cc6f23880c\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.383358 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd9lz\" (UniqueName: \"kubernetes.io/projected/6e5c2028-7096-441e-a33d-60cc6f23880c-kube-api-access-cd9lz\") pod \"6e5c2028-7096-441e-a33d-60cc6f23880c\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.383433 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-dispersionconf\") pod \"6e5c2028-7096-441e-a33d-60cc6f23880c\" (UID: \"6e5c2028-7096-441e-a33d-60cc6f23880c\") " Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.384856 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6e5c2028-7096-441e-a33d-60cc6f23880c" (UID: "6e5c2028-7096-441e-a33d-60cc6f23880c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.386679 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-scripts" (OuterVolumeSpecName: "scripts") pod "6e5c2028-7096-441e-a33d-60cc6f23880c" (UID: "6e5c2028-7096-441e-a33d-60cc6f23880c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.386837 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e5c2028-7096-441e-a33d-60cc6f23880c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6e5c2028-7096-441e-a33d-60cc6f23880c" (UID: "6e5c2028-7096-441e-a33d-60cc6f23880c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.387024 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.387517 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.390007 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e5c2028-7096-441e-a33d-60cc6f23880c" (UID: "6e5c2028-7096-441e-a33d-60cc6f23880c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.391699 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6e5c2028-7096-441e-a33d-60cc6f23880c" (UID: "6e5c2028-7096-441e-a33d-60cc6f23880c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.395124 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6e5c2028-7096-441e-a33d-60cc6f23880c" (UID: "6e5c2028-7096-441e-a33d-60cc6f23880c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.397199 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e5c2028-7096-441e-a33d-60cc6f23880c-kube-api-access-cd9lz" (OuterVolumeSpecName: "kube-api-access-cd9lz") pod "6e5c2028-7096-441e-a33d-60cc6f23880c" (UID: "6e5c2028-7096-441e-a33d-60cc6f23880c"). InnerVolumeSpecName "kube-api-access-cd9lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.487754 4848 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.488836 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.488856 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e5c2028-7096-441e-a33d-60cc6f23880c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.488868 4848 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6e5c2028-7096-441e-a33d-60cc6f23880c-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.488879 4848 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.488892 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd9lz\" (UniqueName: \"kubernetes.io/projected/6e5c2028-7096-441e-a33d-60cc6f23880c-kube-api-access-cd9lz\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.488905 4848 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6e5c2028-7096-441e-a33d-60cc6f23880c-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.861721 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gplgt"] Dec 04 14:10:49 crc kubenswrapper[4848]: I1204 14:10:49.872480 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:10:50 crc kubenswrapper[4848]: I1204 14:10:50.318605 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gplgt" event={"ID":"ff826431-be61-4dea-82e7-b6849d838e52","Type":"ContainerStarted","Data":"913f947594ec9e9c4da5dac9215dd5c77588bb5b2cb747693c4db75904a84b64"} Dec 04 14:10:50 crc kubenswrapper[4848]: I1204 14:10:50.318697 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9ltlr" Dec 04 14:10:50 crc kubenswrapper[4848]: I1204 14:10:50.412633 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9ltlr"] Dec 04 14:10:50 crc kubenswrapper[4848]: I1204 14:10:50.421383 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-9ltlr"] Dec 04 14:10:50 crc kubenswrapper[4848]: I1204 14:10:50.863634 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:50 crc kubenswrapper[4848]: I1204 14:10:50.864008 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:52 crc kubenswrapper[4848]: I1204 14:10:52.054721 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:10:52 crc kubenswrapper[4848]: E1204 14:10:52.054939 4848 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:10:52 crc kubenswrapper[4848]: E1204 14:10:52.054989 4848 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:10:52 crc kubenswrapper[4848]: E1204 14:10:52.055056 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift podName:0102871d-c950-46e7-af43-a2b3f1cbc39f nodeName:}" failed. No retries permitted until 2025-12-04 14:11:00.055038637 +0000 UTC m=+1363.997535165 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift") pod "swift-storage-0" (UID: "0102871d-c950-46e7-af43-a2b3f1cbc39f") : configmap "swift-ring-files" not found Dec 04 14:10:52 crc kubenswrapper[4848]: I1204 14:10:52.406100 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e5c2028-7096-441e-a33d-60cc6f23880c" path="/var/lib/kubelet/pods/6e5c2028-7096-441e-a33d-60cc6f23880c/volumes" Dec 04 14:10:52 crc kubenswrapper[4848]: I1204 14:10:52.433684 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 04 14:10:52 crc kubenswrapper[4848]: I1204 14:10:52.514831 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 04 14:10:52 crc kubenswrapper[4848]: I1204 14:10:52.900570 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 14:10:53 crc kubenswrapper[4848]: I1204 14:10:53.417094 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:10:53 crc kubenswrapper[4848]: I1204 14:10:53.489703 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dpbbw"] Dec 04 14:10:53 crc kubenswrapper[4848]: I1204 14:10:53.489926 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" podUID="f66fc099-8a2e-49ea-803c-417aeedbe389" containerName="dnsmasq-dns" containerID="cri-o://c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012" gracePeriod=10 Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.010915 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:54 crc kubenswrapper[4848]: E1204 14:10:54.096716 4848 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.5:34050->38.102.83.5:35999: write tcp 38.102.83.5:34050->38.102.83.5:35999: write: broken pipe Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.114403 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-dns-svc\") pod \"f66fc099-8a2e-49ea-803c-417aeedbe389\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.114683 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x6qs\" (UniqueName: \"kubernetes.io/projected/f66fc099-8a2e-49ea-803c-417aeedbe389-kube-api-access-2x6qs\") pod \"f66fc099-8a2e-49ea-803c-417aeedbe389\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.114727 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-config\") pod \"f66fc099-8a2e-49ea-803c-417aeedbe389\" (UID: \"f66fc099-8a2e-49ea-803c-417aeedbe389\") " Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.121610 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f66fc099-8a2e-49ea-803c-417aeedbe389-kube-api-access-2x6qs" (OuterVolumeSpecName: "kube-api-access-2x6qs") pod "f66fc099-8a2e-49ea-803c-417aeedbe389" (UID: "f66fc099-8a2e-49ea-803c-417aeedbe389"). InnerVolumeSpecName "kube-api-access-2x6qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.174497 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f66fc099-8a2e-49ea-803c-417aeedbe389" (UID: "f66fc099-8a2e-49ea-803c-417aeedbe389"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.181728 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-config" (OuterVolumeSpecName: "config") pod "f66fc099-8a2e-49ea-803c-417aeedbe389" (UID: "f66fc099-8a2e-49ea-803c-417aeedbe389"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.216927 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x6qs\" (UniqueName: \"kubernetes.io/projected/f66fc099-8a2e-49ea-803c-417aeedbe389-kube-api-access-2x6qs\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.216995 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.217008 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f66fc099-8a2e-49ea-803c-417aeedbe389-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.363724 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerStarted","Data":"57084ce6e036e4bb452a513c26295d75c4c3ba17caf688ccffdc8a807ef9505e"} Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.365793 4848 generic.go:334] "Generic (PLEG): container finished" podID="f66fc099-8a2e-49ea-803c-417aeedbe389" containerID="c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012" exitCode=0 Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.365820 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" event={"ID":"f66fc099-8a2e-49ea-803c-417aeedbe389","Type":"ContainerDied","Data":"c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012"} Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.365842 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" event={"ID":"f66fc099-8a2e-49ea-803c-417aeedbe389","Type":"ContainerDied","Data":"b0fb18610625685b47feb11a9c646e7e48136a5356be9218e7b8bfd710809bc3"} Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.365859 4848 scope.go:117] "RemoveContainer" containerID="c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012" Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.365922 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dpbbw" Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.428836 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dpbbw"] Dec 04 14:10:54 crc kubenswrapper[4848]: I1204 14:10:54.428871 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dpbbw"] Dec 04 14:10:56 crc kubenswrapper[4848]: I1204 14:10:56.344575 4848 scope.go:117] "RemoveContainer" containerID="8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc" Dec 04 14:10:56 crc kubenswrapper[4848]: I1204 14:10:56.406649 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f66fc099-8a2e-49ea-803c-417aeedbe389" path="/var/lib/kubelet/pods/f66fc099-8a2e-49ea-803c-417aeedbe389/volumes" Dec 04 14:10:56 crc kubenswrapper[4848]: I1204 14:10:56.561499 4848 scope.go:117] "RemoveContainer" containerID="c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012" Dec 04 14:10:56 crc kubenswrapper[4848]: E1204 14:10:56.562333 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012\": container with ID starting with c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012 not found: ID does not exist" containerID="c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012" Dec 04 14:10:56 crc kubenswrapper[4848]: I1204 14:10:56.563331 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012"} err="failed to get container status \"c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012\": rpc error: code = NotFound desc = could not find container \"c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012\": container with ID starting with c34b75eb04d6123b861e025a908231a9f121cea62c3f9edbd757d3b52cabc012 not found: ID does not exist" Dec 04 14:10:56 crc kubenswrapper[4848]: I1204 14:10:56.563425 4848 scope.go:117] "RemoveContainer" containerID="8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc" Dec 04 14:10:56 crc kubenswrapper[4848]: E1204 14:10:56.567078 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc\": container with ID starting with 8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc not found: ID does not exist" containerID="8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc" Dec 04 14:10:56 crc kubenswrapper[4848]: I1204 14:10:56.567121 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc"} err="failed to get container status \"8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc\": rpc error: code = NotFound desc = could not find container \"8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc\": container with ID starting with 8ec8f2ae632b8a2ee202ee01534baedd018572f6380cfcfb36106507fffa29dc not found: ID does not exist" Dec 04 14:10:56 crc kubenswrapper[4848]: I1204 14:10:56.988488 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:57 crc kubenswrapper[4848]: I1204 14:10:57.088997 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 04 14:10:57 crc kubenswrapper[4848]: I1204 14:10:57.412281 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gplgt" event={"ID":"ff826431-be61-4dea-82e7-b6849d838e52","Type":"ContainerStarted","Data":"57a3c26e09f627ab760ca9d42d4b4c915494ab0fd00e1915af3600711971a72d"} Dec 04 14:10:57 crc kubenswrapper[4848]: I1204 14:10:57.419069 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerStarted","Data":"4a29902bc73018a50387e59b45b3c5a4eec4d42ef6f1c00e1528e361cd96a8c4"} Dec 04 14:10:57 crc kubenswrapper[4848]: I1204 14:10:57.421453 4848 generic.go:334] "Generic (PLEG): container finished" podID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerID="0f318664aee27c7d28309de918f73ef54729735695edff04d26277043b9fe4de" exitCode=0 Dec 04 14:10:57 crc kubenswrapper[4848]: I1204 14:10:57.423226 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30883c46-53ec-45f9-84f6-29a64ccf8994","Type":"ContainerDied","Data":"0f318664aee27c7d28309de918f73ef54729735695edff04d26277043b9fe4de"} Dec 04 14:10:57 crc kubenswrapper[4848]: I1204 14:10:57.448758 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-gplgt" podStartSLOduration=2.755999099 podStartE2EDuration="9.44874103s" podCreationTimestamp="2025-12-04 14:10:48 +0000 UTC" firstStartedPulling="2025-12-04 14:10:49.872233492 +0000 UTC m=+1353.814730020" lastFinishedPulling="2025-12-04 14:10:56.564975423 +0000 UTC m=+1360.507471951" observedRunningTime="2025-12-04 14:10:57.445659676 +0000 UTC m=+1361.388156224" watchObservedRunningTime="2025-12-04 14:10:57.44874103 +0000 UTC m=+1361.391237558" Dec 04 14:10:58 crc kubenswrapper[4848]: I1204 14:10:58.435071 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30883c46-53ec-45f9-84f6-29a64ccf8994","Type":"ContainerStarted","Data":"bfa96d34d9ae2fa019286f6eeeaf829b92a08bf46a37c957eb9adf61e53823d7"} Dec 04 14:10:58 crc kubenswrapper[4848]: I1204 14:10:58.435699 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:10:58 crc kubenswrapper[4848]: I1204 14:10:58.437248 4848 generic.go:334] "Generic (PLEG): container finished" podID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerID="c4829738b3394f57da24a7580486aa1ca47db4fc2c06fe141440e1c598436e14" exitCode=0 Dec 04 14:10:58 crc kubenswrapper[4848]: I1204 14:10:58.437339 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6049ed43-40db-4bff-b89b-88cce285fd2d","Type":"ContainerDied","Data":"c4829738b3394f57da24a7580486aa1ca47db4fc2c06fe141440e1c598436e14"} Dec 04 14:10:58 crc kubenswrapper[4848]: I1204 14:10:58.466609 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.776847895 podStartE2EDuration="52.466588671s" podCreationTimestamp="2025-12-04 14:10:06 +0000 UTC" firstStartedPulling="2025-12-04 14:10:11.771346747 +0000 UTC m=+1315.713843265" lastFinishedPulling="2025-12-04 14:10:23.461087513 +0000 UTC m=+1327.403584041" observedRunningTime="2025-12-04 14:10:58.454328395 +0000 UTC m=+1362.396824923" watchObservedRunningTime="2025-12-04 14:10:58.466588671 +0000 UTC m=+1362.409085199" Dec 04 14:10:59 crc kubenswrapper[4848]: I1204 14:10:59.451036 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6049ed43-40db-4bff-b89b-88cce285fd2d","Type":"ContainerStarted","Data":"11db54d0028a98c6804adef44dccc01f2bbd27f43a15550ce6f0f8f2a8af3d45"} Dec 04 14:10:59 crc kubenswrapper[4848]: I1204 14:10:59.453392 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rg7xl" event={"ID":"ef9a0101-72b3-4879-a93b-f58362d175ed","Type":"ContainerStarted","Data":"e48b15a5e7fe982feb371c2a11d0366b88bde2a2c709bcadc3ad352fd8a23f59"} Dec 04 14:10:59 crc kubenswrapper[4848]: I1204 14:10:59.484061 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rg7xl" podStartSLOduration=12.686465663 podStartE2EDuration="44.484043809s" podCreationTimestamp="2025-12-04 14:10:15 +0000 UTC" firstStartedPulling="2025-12-04 14:10:26.19011235 +0000 UTC m=+1330.132608868" lastFinishedPulling="2025-12-04 14:10:57.987690466 +0000 UTC m=+1361.930187014" observedRunningTime="2025-12-04 14:10:59.479453208 +0000 UTC m=+1363.421949766" watchObservedRunningTime="2025-12-04 14:10:59.484043809 +0000 UTC m=+1363.426540327" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.152025 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:11:00 crc kubenswrapper[4848]: E1204 14:11:00.152229 4848 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 14:11:00 crc kubenswrapper[4848]: E1204 14:11:00.152501 4848 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 14:11:00 crc kubenswrapper[4848]: E1204 14:11:00.152561 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift podName:0102871d-c950-46e7-af43-a2b3f1cbc39f nodeName:}" failed. No retries permitted until 2025-12-04 14:11:16.152540662 +0000 UTC m=+1380.095037200 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift") pod "swift-storage-0" (UID: "0102871d-c950-46e7-af43-a2b3f1cbc39f") : configmap "swift-ring-files" not found Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.203245 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-8554695464-z4rxq" podUID="457b813b-6604-460c-b16a-64ea6bb703fc" containerName="console" containerID="cri-o://febbc6b9ab66de76daf5f91b11e58fc53fe3646931b34056207206bb283c64b0" gracePeriod=15 Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.467340 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8554695464-z4rxq_457b813b-6604-460c-b16a-64ea6bb703fc/console/0.log" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.467392 4848 generic.go:334] "Generic (PLEG): container finished" podID="457b813b-6604-460c-b16a-64ea6bb703fc" containerID="febbc6b9ab66de76daf5f91b11e58fc53fe3646931b34056207206bb283c64b0" exitCode=2 Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.467504 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8554695464-z4rxq" event={"ID":"457b813b-6604-460c-b16a-64ea6bb703fc","Type":"ContainerDied","Data":"febbc6b9ab66de76daf5f91b11e58fc53fe3646931b34056207206bb283c64b0"} Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.467587 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.498603 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.498573616 podStartE2EDuration="54.498573616s" podCreationTimestamp="2025-12-04 14:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:00.489028604 +0000 UTC m=+1364.431525132" watchObservedRunningTime="2025-12-04 14:11:00.498573616 +0000 UTC m=+1364.441070144" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.643555 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-99d5-account-create-update-j5z7r"] Dec 04 14:11:00 crc kubenswrapper[4848]: E1204 14:11:00.643919 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66fc099-8a2e-49ea-803c-417aeedbe389" containerName="init" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.643935 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66fc099-8a2e-49ea-803c-417aeedbe389" containerName="init" Dec 04 14:11:00 crc kubenswrapper[4848]: E1204 14:11:00.643979 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66fc099-8a2e-49ea-803c-417aeedbe389" containerName="dnsmasq-dns" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.643986 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66fc099-8a2e-49ea-803c-417aeedbe389" containerName="dnsmasq-dns" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.644164 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66fc099-8a2e-49ea-803c-417aeedbe389" containerName="dnsmasq-dns" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.644934 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.647496 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.659715 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-99d5-account-create-update-j5z7r"] Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.726044 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2z5jg"] Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.727697 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.735403 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2z5jg"] Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.762175 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-operator-scripts\") pod \"keystone-99d5-account-create-update-j5z7r\" (UID: \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\") " pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.762370 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmlxs\" (UniqueName: \"kubernetes.io/projected/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-kube-api-access-kmlxs\") pod \"keystone-99d5-account-create-update-j5z7r\" (UID: \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\") " pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.864526 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmlxs\" (UniqueName: \"kubernetes.io/projected/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-kube-api-access-kmlxs\") pod \"keystone-99d5-account-create-update-j5z7r\" (UID: \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\") " pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.864633 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-operator-scripts\") pod \"keystone-99d5-account-create-update-j5z7r\" (UID: \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\") " pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.864658 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rjsd\" (UniqueName: \"kubernetes.io/projected/a530fe44-72a5-487a-8b28-fae47fa307c5-kube-api-access-5rjsd\") pod \"keystone-db-create-2z5jg\" (UID: \"a530fe44-72a5-487a-8b28-fae47fa307c5\") " pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.864691 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a530fe44-72a5-487a-8b28-fae47fa307c5-operator-scripts\") pod \"keystone-db-create-2z5jg\" (UID: \"a530fe44-72a5-487a-8b28-fae47fa307c5\") " pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.865329 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-operator-scripts\") pod \"keystone-99d5-account-create-update-j5z7r\" (UID: \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\") " pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.883538 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmlxs\" (UniqueName: \"kubernetes.io/projected/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-kube-api-access-kmlxs\") pod \"keystone-99d5-account-create-update-j5z7r\" (UID: \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\") " pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.919054 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-mk75p"] Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.920656 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mk75p" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.921145 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-rg7xl" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.927770 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mk75p"] Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.966187 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.966968 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rjsd\" (UniqueName: \"kubernetes.io/projected/a530fe44-72a5-487a-8b28-fae47fa307c5-kube-api-access-5rjsd\") pod \"keystone-db-create-2z5jg\" (UID: \"a530fe44-72a5-487a-8b28-fae47fa307c5\") " pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.967025 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a530fe44-72a5-487a-8b28-fae47fa307c5-operator-scripts\") pod \"keystone-db-create-2z5jg\" (UID: \"a530fe44-72a5-487a-8b28-fae47fa307c5\") " pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:00 crc kubenswrapper[4848]: I1204 14:11:00.968827 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a530fe44-72a5-487a-8b28-fae47fa307c5-operator-scripts\") pod \"keystone-db-create-2z5jg\" (UID: \"a530fe44-72a5-487a-8b28-fae47fa307c5\") " pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:00.999510 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rjsd\" (UniqueName: \"kubernetes.io/projected/a530fe44-72a5-487a-8b28-fae47fa307c5-kube-api-access-5rjsd\") pod \"keystone-db-create-2z5jg\" (UID: \"a530fe44-72a5-487a-8b28-fae47fa307c5\") " pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.038705 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-47f3-account-create-update-sk9vk"] Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.040798 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.045634 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-47f3-account-create-update-sk9vk"] Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.046072 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.053415 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.070902 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvczq\" (UniqueName: \"kubernetes.io/projected/6fa01532-be1d-4542-bad2-3608980a8835-kube-api-access-zvczq\") pod \"placement-db-create-mk75p\" (UID: \"6fa01532-be1d-4542-bad2-3608980a8835\") " pod="openstack/placement-db-create-mk75p" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.070988 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fa01532-be1d-4542-bad2-3608980a8835-operator-scripts\") pod \"placement-db-create-mk75p\" (UID: \"6fa01532-be1d-4542-bad2-3608980a8835\") " pod="openstack/placement-db-create-mk75p" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.185106 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvczq\" (UniqueName: \"kubernetes.io/projected/6fa01532-be1d-4542-bad2-3608980a8835-kube-api-access-zvczq\") pod \"placement-db-create-mk75p\" (UID: \"6fa01532-be1d-4542-bad2-3608980a8835\") " pod="openstack/placement-db-create-mk75p" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.185156 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fa01532-be1d-4542-bad2-3608980a8835-operator-scripts\") pod \"placement-db-create-mk75p\" (UID: \"6fa01532-be1d-4542-bad2-3608980a8835\") " pod="openstack/placement-db-create-mk75p" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.185235 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnzrx\" (UniqueName: \"kubernetes.io/projected/d6dcffe4-90e2-4334-88c9-60b336d87402-kube-api-access-bnzrx\") pod \"placement-47f3-account-create-update-sk9vk\" (UID: \"d6dcffe4-90e2-4334-88c9-60b336d87402\") " pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.185290 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6dcffe4-90e2-4334-88c9-60b336d87402-operator-scripts\") pod \"placement-47f3-account-create-update-sk9vk\" (UID: \"d6dcffe4-90e2-4334-88c9-60b336d87402\") " pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.186495 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fa01532-be1d-4542-bad2-3608980a8835-operator-scripts\") pod \"placement-db-create-mk75p\" (UID: \"6fa01532-be1d-4542-bad2-3608980a8835\") " pod="openstack/placement-db-create-mk75p" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.206561 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvczq\" (UniqueName: \"kubernetes.io/projected/6fa01532-be1d-4542-bad2-3608980a8835-kube-api-access-zvczq\") pod \"placement-db-create-mk75p\" (UID: \"6fa01532-be1d-4542-bad2-3608980a8835\") " pod="openstack/placement-db-create-mk75p" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.261037 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-qbr2v"] Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.268584 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.283595 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-qbr2v"] Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.288343 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mk75p" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.294386 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnzrx\" (UniqueName: \"kubernetes.io/projected/d6dcffe4-90e2-4334-88c9-60b336d87402-kube-api-access-bnzrx\") pod \"placement-47f3-account-create-update-sk9vk\" (UID: \"d6dcffe4-90e2-4334-88c9-60b336d87402\") " pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.294455 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6dcffe4-90e2-4334-88c9-60b336d87402-operator-scripts\") pod \"placement-47f3-account-create-update-sk9vk\" (UID: \"d6dcffe4-90e2-4334-88c9-60b336d87402\") " pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.295348 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6dcffe4-90e2-4334-88c9-60b336d87402-operator-scripts\") pod \"placement-47f3-account-create-update-sk9vk\" (UID: \"d6dcffe4-90e2-4334-88c9-60b336d87402\") " pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.326153 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnzrx\" (UniqueName: \"kubernetes.io/projected/d6dcffe4-90e2-4334-88c9-60b336d87402-kube-api-access-bnzrx\") pod \"placement-47f3-account-create-update-sk9vk\" (UID: \"d6dcffe4-90e2-4334-88c9-60b336d87402\") " pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.369342 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.370704 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8554695464-z4rxq_457b813b-6604-460c-b16a-64ea6bb703fc/console/0.log" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.370759 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.371141 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-0e75-account-create-update-x7wmg"] Dec 04 14:11:01 crc kubenswrapper[4848]: E1204 14:11:01.371472 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457b813b-6604-460c-b16a-64ea6bb703fc" containerName="console" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.371484 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="457b813b-6604-460c-b16a-64ea6bb703fc" containerName="console" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.371691 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="457b813b-6604-460c-b16a-64ea6bb703fc" containerName="console" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.372316 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.374506 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.386604 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0e75-account-create-update-x7wmg"] Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.397466 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5592df9-94ea-4dd6-9f76-473b02a44897-operator-scripts\") pod \"glance-db-create-qbr2v\" (UID: \"b5592df9-94ea-4dd6-9f76-473b02a44897\") " pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.397596 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24795\" (UniqueName: \"kubernetes.io/projected/b5592df9-94ea-4dd6-9f76-473b02a44897-kube-api-access-24795\") pod \"glance-db-create-qbr2v\" (UID: \"b5592df9-94ea-4dd6-9f76-473b02a44897\") " pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.487068 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerStarted","Data":"14f013eceadc2ad0e7111f2fe485eefe5cb5a76b319302c45f850d5966f830ea"} Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.494465 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8554695464-z4rxq_457b813b-6604-460c-b16a-64ea6bb703fc/console/0.log" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.494936 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8554695464-z4rxq" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.495347 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8554695464-z4rxq" event={"ID":"457b813b-6604-460c-b16a-64ea6bb703fc","Type":"ContainerDied","Data":"32611b84f5c8115d7d9cd15f76b2e89d102752cc06249df92c85dfc4197b367d"} Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.495382 4848 scope.go:117] "RemoveContainer" containerID="febbc6b9ab66de76daf5f91b11e58fc53fe3646931b34056207206bb283c64b0" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499045 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-service-ca\") pod \"457b813b-6604-460c-b16a-64ea6bb703fc\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499160 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rds62\" (UniqueName: \"kubernetes.io/projected/457b813b-6604-460c-b16a-64ea6bb703fc-kube-api-access-rds62\") pod \"457b813b-6604-460c-b16a-64ea6bb703fc\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499198 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-oauth-serving-cert\") pod \"457b813b-6604-460c-b16a-64ea6bb703fc\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499286 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-oauth-config\") pod \"457b813b-6604-460c-b16a-64ea6bb703fc\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499336 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-serving-cert\") pod \"457b813b-6604-460c-b16a-64ea6bb703fc\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499378 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-trusted-ca-bundle\") pod \"457b813b-6604-460c-b16a-64ea6bb703fc\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499409 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-console-config\") pod \"457b813b-6604-460c-b16a-64ea6bb703fc\" (UID: \"457b813b-6604-460c-b16a-64ea6bb703fc\") " Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499815 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5592df9-94ea-4dd6-9f76-473b02a44897-operator-scripts\") pod \"glance-db-create-qbr2v\" (UID: \"b5592df9-94ea-4dd6-9f76-473b02a44897\") " pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499849 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24795\" (UniqueName: \"kubernetes.io/projected/b5592df9-94ea-4dd6-9f76-473b02a44897-kube-api-access-24795\") pod \"glance-db-create-qbr2v\" (UID: \"b5592df9-94ea-4dd6-9f76-473b02a44897\") " pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.499930 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa26501-8eec-4719-9b4d-211dc1e24ec9-operator-scripts\") pod \"glance-0e75-account-create-update-x7wmg\" (UID: \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\") " pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.500038 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4lq5\" (UniqueName: \"kubernetes.io/projected/2fa26501-8eec-4719-9b4d-211dc1e24ec9-kube-api-access-f4lq5\") pod \"glance-0e75-account-create-update-x7wmg\" (UID: \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\") " pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.501206 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-console-config" (OuterVolumeSpecName: "console-config") pod "457b813b-6604-460c-b16a-64ea6bb703fc" (UID: "457b813b-6604-460c-b16a-64ea6bb703fc"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.502034 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-service-ca" (OuterVolumeSpecName: "service-ca") pod "457b813b-6604-460c-b16a-64ea6bb703fc" (UID: "457b813b-6604-460c-b16a-64ea6bb703fc"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.504741 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "457b813b-6604-460c-b16a-64ea6bb703fc" (UID: "457b813b-6604-460c-b16a-64ea6bb703fc"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.505579 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5592df9-94ea-4dd6-9f76-473b02a44897-operator-scripts\") pod \"glance-db-create-qbr2v\" (UID: \"b5592df9-94ea-4dd6-9f76-473b02a44897\") " pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.505651 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "457b813b-6604-460c-b16a-64ea6bb703fc" (UID: "457b813b-6604-460c-b16a-64ea6bb703fc"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.506628 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/457b813b-6604-460c-b16a-64ea6bb703fc-kube-api-access-rds62" (OuterVolumeSpecName: "kube-api-access-rds62") pod "457b813b-6604-460c-b16a-64ea6bb703fc" (UID: "457b813b-6604-460c-b16a-64ea6bb703fc"). InnerVolumeSpecName "kube-api-access-rds62". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.506705 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "457b813b-6604-460c-b16a-64ea6bb703fc" (UID: "457b813b-6604-460c-b16a-64ea6bb703fc"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.506719 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "457b813b-6604-460c-b16a-64ea6bb703fc" (UID: "457b813b-6604-460c-b16a-64ea6bb703fc"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.515692 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=13.052897626 podStartE2EDuration="49.515668495s" podCreationTimestamp="2025-12-04 14:10:12 +0000 UTC" firstStartedPulling="2025-12-04 14:10:24.624637308 +0000 UTC m=+1328.567133856" lastFinishedPulling="2025-12-04 14:11:01.087408207 +0000 UTC m=+1365.029904725" observedRunningTime="2025-12-04 14:11:01.511988855 +0000 UTC m=+1365.454485383" watchObservedRunningTime="2025-12-04 14:11:01.515668495 +0000 UTC m=+1365.458165023" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.520257 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24795\" (UniqueName: \"kubernetes.io/projected/b5592df9-94ea-4dd6-9f76-473b02a44897-kube-api-access-24795\") pod \"glance-db-create-qbr2v\" (UID: \"b5592df9-94ea-4dd6-9f76-473b02a44897\") " pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603295 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa26501-8eec-4719-9b4d-211dc1e24ec9-operator-scripts\") pod \"glance-0e75-account-create-update-x7wmg\" (UID: \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\") " pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603355 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4lq5\" (UniqueName: \"kubernetes.io/projected/2fa26501-8eec-4719-9b4d-211dc1e24ec9-kube-api-access-f4lq5\") pod \"glance-0e75-account-create-update-x7wmg\" (UID: \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\") " pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603443 4848 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603454 4848 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/457b813b-6604-460c-b16a-64ea6bb703fc-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603466 4848 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603476 4848 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603484 4848 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603492 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rds62\" (UniqueName: \"kubernetes.io/projected/457b813b-6604-460c-b16a-64ea6bb703fc-kube-api-access-rds62\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.603502 4848 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/457b813b-6604-460c-b16a-64ea6bb703fc-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.605660 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa26501-8eec-4719-9b4d-211dc1e24ec9-operator-scripts\") pod \"glance-0e75-account-create-update-x7wmg\" (UID: \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\") " pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.608386 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.622393 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4lq5\" (UniqueName: \"kubernetes.io/projected/2fa26501-8eec-4719-9b4d-211dc1e24ec9-kube-api-access-f4lq5\") pod \"glance-0e75-account-create-update-x7wmg\" (UID: \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\") " pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.680052 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-99d5-account-create-update-j5z7r"] Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.687727 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2z5jg"] Dec 04 14:11:01 crc kubenswrapper[4848]: W1204 14:11:01.688621 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda530fe44_72a5_487a_8b28_fae47fa307c5.slice/crio-02cc77d804a356dff77333f5c4aeb8f3c33e7ba8699cd90156137a898ade4ee6 WatchSource:0}: Error finding container 02cc77d804a356dff77333f5c4aeb8f3c33e7ba8699cd90156137a898ade4ee6: Status 404 returned error can't find the container with id 02cc77d804a356dff77333f5c4aeb8f3c33e7ba8699cd90156137a898ade4ee6 Dec 04 14:11:01 crc kubenswrapper[4848]: W1204 14:11:01.694877 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded0bfa2e_a25b_4dce_86be_1401ac97f10a.slice/crio-985244c746bc5955b1eb02e933b3dca30653ef7d644d5416e6e19bcd0438bda2 WatchSource:0}: Error finding container 985244c746bc5955b1eb02e933b3dca30653ef7d644d5416e6e19bcd0438bda2: Status 404 returned error can't find the container with id 985244c746bc5955b1eb02e933b3dca30653ef7d644d5416e6e19bcd0438bda2 Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.696811 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:01 crc kubenswrapper[4848]: I1204 14:11:01.941328 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mk75p"] Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.013762 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-8554695464-z4rxq"] Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.026234 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-8554695464-z4rxq"] Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.053060 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-47f3-account-create-update-sk9vk"] Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.138182 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-qbr2v"] Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.333982 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0e75-account-create-update-x7wmg"] Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.405862 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="457b813b-6604-460c-b16a-64ea6bb703fc" path="/var/lib/kubelet/pods/457b813b-6604-460c-b16a-64ea6bb703fc/volumes" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.502758 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0e75-account-create-update-x7wmg" event={"ID":"2fa26501-8eec-4719-9b4d-211dc1e24ec9","Type":"ContainerStarted","Data":"2f8674b94f10deef94971203c1ccfd0868cd2be01ba08373f53c5613068cf91f"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.505520 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-47f3-account-create-update-sk9vk" event={"ID":"d6dcffe4-90e2-4334-88c9-60b336d87402","Type":"ContainerStarted","Data":"93a0f28489a9a7797f4f30b900479579fb3e7642d7dc0600f93932c052ddb1b3"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.505551 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-47f3-account-create-update-sk9vk" event={"ID":"d6dcffe4-90e2-4334-88c9-60b336d87402","Type":"ContainerStarted","Data":"b3bba2c1493986393ddd02b557cf6e9e08a325ec8515203571982d04cddd3536"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.507222 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mk75p" event={"ID":"6fa01532-be1d-4542-bad2-3608980a8835","Type":"ContainerStarted","Data":"33a63e61f1aa5d7c9f1803584c7c970dfb19df092c2867dc4f89ecdff5877778"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.507247 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mk75p" event={"ID":"6fa01532-be1d-4542-bad2-3608980a8835","Type":"ContainerStarted","Data":"638a5a72c7c213261445134dc581b98f1dbf478ea16dd73e2ab2f9845497b406"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.512981 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qbr2v" event={"ID":"b5592df9-94ea-4dd6-9f76-473b02a44897","Type":"ContainerStarted","Data":"908aba6352453cf2813c6304c3364f57f748e082b0c5543416eded9c86f227af"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.518809 4848 generic.go:334] "Generic (PLEG): container finished" podID="a530fe44-72a5-487a-8b28-fae47fa307c5" containerID="af28b07da149fcebf8836123adc7855c4bfeff25ee48c659b5044cbd684ad376" exitCode=0 Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.518911 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2z5jg" event={"ID":"a530fe44-72a5-487a-8b28-fae47fa307c5","Type":"ContainerDied","Data":"af28b07da149fcebf8836123adc7855c4bfeff25ee48c659b5044cbd684ad376"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.518934 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2z5jg" event={"ID":"a530fe44-72a5-487a-8b28-fae47fa307c5","Type":"ContainerStarted","Data":"02cc77d804a356dff77333f5c4aeb8f3c33e7ba8699cd90156137a898ade4ee6"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.523773 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-99d5-account-create-update-j5z7r" event={"ID":"ed0bfa2e-a25b-4dce-86be-1401ac97f10a","Type":"ContainerStarted","Data":"a2cbfeeeaeff9ac2880d140d734bc5830537af1f4d7bfccbf22297c6650b0925"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.523811 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-99d5-account-create-update-j5z7r" event={"ID":"ed0bfa2e-a25b-4dce-86be-1401ac97f10a","Type":"ContainerStarted","Data":"985244c746bc5955b1eb02e933b3dca30653ef7d644d5416e6e19bcd0438bda2"} Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.537980 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-47f3-account-create-update-sk9vk" podStartSLOduration=1.537962899 podStartE2EDuration="1.537962899s" podCreationTimestamp="2025-12-04 14:11:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:02.533222315 +0000 UTC m=+1366.475718833" watchObservedRunningTime="2025-12-04 14:11:02.537962899 +0000 UTC m=+1366.480459417" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.602023 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-mk75p" podStartSLOduration=2.602003073 podStartE2EDuration="2.602003073s" podCreationTimestamp="2025-12-04 14:11:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:02.574844393 +0000 UTC m=+1366.517340921" watchObservedRunningTime="2025-12-04 14:11:02.602003073 +0000 UTC m=+1366.544499611" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.618196 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-99d5-account-create-update-j5z7r" podStartSLOduration=2.618173224 podStartE2EDuration="2.618173224s" podCreationTimestamp="2025-12-04 14:11:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:02.615408127 +0000 UTC m=+1366.557904655" watchObservedRunningTime="2025-12-04 14:11:02.618173224 +0000 UTC m=+1366.560669752" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.790301 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-cgjg9"] Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.791869 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.841882 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-cgjg9\" (UID: \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\") " pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.842077 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89rtf\" (UniqueName: \"kubernetes.io/projected/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-kube-api-access-89rtf\") pod \"mysqld-exporter-openstack-db-create-cgjg9\" (UID: \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\") " pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.842239 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-cgjg9"] Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.944103 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-cgjg9\" (UID: \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\") " pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.944189 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89rtf\" (UniqueName: \"kubernetes.io/projected/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-kube-api-access-89rtf\") pod \"mysqld-exporter-openstack-db-create-cgjg9\" (UID: \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\") " pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.947774 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-cgjg9\" (UID: \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\") " pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:02 crc kubenswrapper[4848]: I1204 14:11:02.998723 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89rtf\" (UniqueName: \"kubernetes.io/projected/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-kube-api-access-89rtf\") pod \"mysqld-exporter-openstack-db-create-cgjg9\" (UID: \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\") " pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.028375 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-fd6f-account-create-update-wwnqm"] Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.037366 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.043270 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.048024 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-fd6f-account-create-update-wwnqm"] Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.116424 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.147900 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlfk8\" (UniqueName: \"kubernetes.io/projected/7e78126d-0db0-4fee-a1b6-c23818909b5a-kube-api-access-qlfk8\") pod \"mysqld-exporter-fd6f-account-create-update-wwnqm\" (UID: \"7e78126d-0db0-4fee-a1b6-c23818909b5a\") " pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.148132 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e78126d-0db0-4fee-a1b6-c23818909b5a-operator-scripts\") pod \"mysqld-exporter-fd6f-account-create-update-wwnqm\" (UID: \"7e78126d-0db0-4fee-a1b6-c23818909b5a\") " pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.250183 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlfk8\" (UniqueName: \"kubernetes.io/projected/7e78126d-0db0-4fee-a1b6-c23818909b5a-kube-api-access-qlfk8\") pod \"mysqld-exporter-fd6f-account-create-update-wwnqm\" (UID: \"7e78126d-0db0-4fee-a1b6-c23818909b5a\") " pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.250601 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e78126d-0db0-4fee-a1b6-c23818909b5a-operator-scripts\") pod \"mysqld-exporter-fd6f-account-create-update-wwnqm\" (UID: \"7e78126d-0db0-4fee-a1b6-c23818909b5a\") " pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.251340 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e78126d-0db0-4fee-a1b6-c23818909b5a-operator-scripts\") pod \"mysqld-exporter-fd6f-account-create-update-wwnqm\" (UID: \"7e78126d-0db0-4fee-a1b6-c23818909b5a\") " pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.279815 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlfk8\" (UniqueName: \"kubernetes.io/projected/7e78126d-0db0-4fee-a1b6-c23818909b5a-kube-api-access-qlfk8\") pod \"mysqld-exporter-fd6f-account-create-update-wwnqm\" (UID: \"7e78126d-0db0-4fee-a1b6-c23818909b5a\") " pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.355257 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.541293 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"97bf1fd2-3a36-418d-826a-6dc60889b70c","Type":"ContainerStarted","Data":"25ca55f9e24b39a1e04c7942c023e57389e23bfb55e2105cbbc538f261046085"} Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.548653 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0e75-account-create-update-x7wmg" event={"ID":"2fa26501-8eec-4719-9b4d-211dc1e24ec9","Type":"ContainerStarted","Data":"cd7bacadcb8cda904706d8879ccf0127bc1caf24053c6067a73e9cf2865a670b"} Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.552058 4848 generic.go:334] "Generic (PLEG): container finished" podID="6fa01532-be1d-4542-bad2-3608980a8835" containerID="33a63e61f1aa5d7c9f1803584c7c970dfb19df092c2867dc4f89ecdff5877778" exitCode=0 Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.552383 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mk75p" event={"ID":"6fa01532-be1d-4542-bad2-3608980a8835","Type":"ContainerDied","Data":"33a63e61f1aa5d7c9f1803584c7c970dfb19df092c2867dc4f89ecdff5877778"} Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.572561 4848 generic.go:334] "Generic (PLEG): container finished" podID="b5592df9-94ea-4dd6-9f76-473b02a44897" containerID="78d3f2e2146a537afba195dca6ac4e0413f68486d7cbd5ac0bb941406bfdff24" exitCode=0 Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.573564 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qbr2v" event={"ID":"b5592df9-94ea-4dd6-9f76-473b02a44897","Type":"ContainerDied","Data":"78d3f2e2146a537afba195dca6ac4e0413f68486d7cbd5ac0bb941406bfdff24"} Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.581404 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=12.035096125 podStartE2EDuration="44.581386816s" podCreationTimestamp="2025-12-04 14:10:19 +0000 UTC" firstStartedPulling="2025-12-04 14:10:30.569649357 +0000 UTC m=+1334.512145885" lastFinishedPulling="2025-12-04 14:11:03.115940048 +0000 UTC m=+1367.058436576" observedRunningTime="2025-12-04 14:11:03.566577127 +0000 UTC m=+1367.509073665" watchObservedRunningTime="2025-12-04 14:11:03.581386816 +0000 UTC m=+1367.523883344" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.629103 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-0e75-account-create-update-x7wmg" podStartSLOduration=2.629077753 podStartE2EDuration="2.629077753s" podCreationTimestamp="2025-12-04 14:11:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:03.616071558 +0000 UTC m=+1367.558568106" watchObservedRunningTime="2025-12-04 14:11:03.629077753 +0000 UTC m=+1367.571574291" Dec 04 14:11:03 crc kubenswrapper[4848]: I1204 14:11:03.664130 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-cgjg9"] Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:03.855810 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-fd6f-account-create-update-wwnqm"] Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.069385 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.219004 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a530fe44-72a5-487a-8b28-fae47fa307c5-operator-scripts\") pod \"a530fe44-72a5-487a-8b28-fae47fa307c5\" (UID: \"a530fe44-72a5-487a-8b28-fae47fa307c5\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.219318 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rjsd\" (UniqueName: \"kubernetes.io/projected/a530fe44-72a5-487a-8b28-fae47fa307c5-kube-api-access-5rjsd\") pod \"a530fe44-72a5-487a-8b28-fae47fa307c5\" (UID: \"a530fe44-72a5-487a-8b28-fae47fa307c5\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.219865 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a530fe44-72a5-487a-8b28-fae47fa307c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a530fe44-72a5-487a-8b28-fae47fa307c5" (UID: "a530fe44-72a5-487a-8b28-fae47fa307c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.222994 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a530fe44-72a5-487a-8b28-fae47fa307c5-kube-api-access-5rjsd" (OuterVolumeSpecName: "kube-api-access-5rjsd") pod "a530fe44-72a5-487a-8b28-fae47fa307c5" (UID: "a530fe44-72a5-487a-8b28-fae47fa307c5"). InnerVolumeSpecName "kube-api-access-5rjsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.322097 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a530fe44-72a5-487a-8b28-fae47fa307c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.322142 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rjsd\" (UniqueName: \"kubernetes.io/projected/a530fe44-72a5-487a-8b28-fae47fa307c5-kube-api-access-5rjsd\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.471006 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.582421 4848 generic.go:334] "Generic (PLEG): container finished" podID="ed0bfa2e-a25b-4dce-86be-1401ac97f10a" containerID="a2cbfeeeaeff9ac2880d140d734bc5830537af1f4d7bfccbf22297c6650b0925" exitCode=0 Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.582459 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-99d5-account-create-update-j5z7r" event={"ID":"ed0bfa2e-a25b-4dce-86be-1401ac97f10a","Type":"ContainerDied","Data":"a2cbfeeeaeff9ac2880d140d734bc5830537af1f4d7bfccbf22297c6650b0925"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.585527 4848 generic.go:334] "Generic (PLEG): container finished" podID="2fa26501-8eec-4719-9b4d-211dc1e24ec9" containerID="cd7bacadcb8cda904706d8879ccf0127bc1caf24053c6067a73e9cf2865a670b" exitCode=0 Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.585558 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0e75-account-create-update-x7wmg" event={"ID":"2fa26501-8eec-4719-9b4d-211dc1e24ec9","Type":"ContainerDied","Data":"cd7bacadcb8cda904706d8879ccf0127bc1caf24053c6067a73e9cf2865a670b"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.587190 4848 generic.go:334] "Generic (PLEG): container finished" podID="d6dcffe4-90e2-4334-88c9-60b336d87402" containerID="93a0f28489a9a7797f4f30b900479579fb3e7642d7dc0600f93932c052ddb1b3" exitCode=0 Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.587253 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-47f3-account-create-update-sk9vk" event={"ID":"d6dcffe4-90e2-4334-88c9-60b336d87402","Type":"ContainerDied","Data":"93a0f28489a9a7797f4f30b900479579fb3e7642d7dc0600f93932c052ddb1b3"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.588989 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"12b7748e-3169-49a0-a634-8a485f7690f4","Type":"ContainerStarted","Data":"5f7d2a3f2940084530b343b19fb1ce0762118b917dba54f69bdb4f310888ff53"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.591316 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" event={"ID":"7e78126d-0db0-4fee-a1b6-c23818909b5a","Type":"ContainerStarted","Data":"34ef3765492d8cb8d0596a210dc5055bb9e57f155788c6e279aa6c369ae7712c"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.591347 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" event={"ID":"7e78126d-0db0-4fee-a1b6-c23818909b5a","Type":"ContainerStarted","Data":"b2f442bafab8c1b12a4224b531b9c9c36961ac391fdbd4e75ee9d25c23be6922"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.596289 4848 generic.go:334] "Generic (PLEG): container finished" podID="bb9c5d54-6432-4f78-a4a1-9dc43ffff102" containerID="777c333b317427bc294a1a4ce5830cf894f0f197a66027b0b4c274b2dc9d7ff9" exitCode=0 Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.596413 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" event={"ID":"bb9c5d54-6432-4f78-a4a1-9dc43ffff102","Type":"ContainerDied","Data":"777c333b317427bc294a1a4ce5830cf894f0f197a66027b0b4c274b2dc9d7ff9"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.596436 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" event={"ID":"bb9c5d54-6432-4f78-a4a1-9dc43ffff102","Type":"ContainerStarted","Data":"2fdc7168af0d3de69a5ccb3941640bacf893e00ff017d536c9ed75695eb1b2f7"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.597930 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2z5jg" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.598616 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2z5jg" event={"ID":"a530fe44-72a5-487a-8b28-fae47fa307c5","Type":"ContainerDied","Data":"02cc77d804a356dff77333f5c4aeb8f3c33e7ba8699cd90156137a898ade4ee6"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.598632 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02cc77d804a356dff77333f5c4aeb8f3c33e7ba8699cd90156137a898ade4ee6" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.673492 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" podStartSLOduration=2.673472364 podStartE2EDuration="2.673472364s" podCreationTimestamp="2025-12-04 14:11:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:04.66342052 +0000 UTC m=+1368.605917048" watchObservedRunningTime="2025-12-04 14:11:04.673472364 +0000 UTC m=+1368.615968892" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:04.708562 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.081214011 podStartE2EDuration="48.708539524s" podCreationTimestamp="2025-12-04 14:10:16 +0000 UTC" firstStartedPulling="2025-12-04 14:10:26.200093631 +0000 UTC m=+1330.142590159" lastFinishedPulling="2025-12-04 14:11:03.827419144 +0000 UTC m=+1367.769915672" observedRunningTime="2025-12-04 14:11:04.697260021 +0000 UTC m=+1368.639756549" watchObservedRunningTime="2025-12-04 14:11:04.708539524 +0000 UTC m=+1368.651036052" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:05.422582 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:05.422738 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:05.608354 4848 generic.go:334] "Generic (PLEG): container finished" podID="ff826431-be61-4dea-82e7-b6849d838e52" containerID="57a3c26e09f627ab760ca9d42d4b4c915494ab0fd00e1915af3600711971a72d" exitCode=0 Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:05.608395 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gplgt" event={"ID":"ff826431-be61-4dea-82e7-b6849d838e52","Type":"ContainerDied","Data":"57a3c26e09f627ab760ca9d42d4b4c915494ab0fd00e1915af3600711971a72d"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:05.610049 4848 generic.go:334] "Generic (PLEG): container finished" podID="7e78126d-0db0-4fee-a1b6-c23818909b5a" containerID="34ef3765492d8cb8d0596a210dc5055bb9e57f155788c6e279aa6c369ae7712c" exitCode=0 Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:05.610132 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" event={"ID":"7e78126d-0db0-4fee-a1b6-c23818909b5a","Type":"ContainerDied","Data":"34ef3765492d8cb8d0596a210dc5055bb9e57f155788c6e279aa6c369ae7712c"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.286560 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.366512 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.373484 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mk75p" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.529463 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvczq\" (UniqueName: \"kubernetes.io/projected/6fa01532-be1d-4542-bad2-3608980a8835-kube-api-access-zvczq\") pod \"6fa01532-be1d-4542-bad2-3608980a8835\" (UID: \"6fa01532-be1d-4542-bad2-3608980a8835\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.529576 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24795\" (UniqueName: \"kubernetes.io/projected/b5592df9-94ea-4dd6-9f76-473b02a44897-kube-api-access-24795\") pod \"b5592df9-94ea-4dd6-9f76-473b02a44897\" (UID: \"b5592df9-94ea-4dd6-9f76-473b02a44897\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.529605 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5592df9-94ea-4dd6-9f76-473b02a44897-operator-scripts\") pod \"b5592df9-94ea-4dd6-9f76-473b02a44897\" (UID: \"b5592df9-94ea-4dd6-9f76-473b02a44897\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.529648 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fa01532-be1d-4542-bad2-3608980a8835-operator-scripts\") pod \"6fa01532-be1d-4542-bad2-3608980a8835\" (UID: \"6fa01532-be1d-4542-bad2-3608980a8835\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.531485 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fa01532-be1d-4542-bad2-3608980a8835-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6fa01532-be1d-4542-bad2-3608980a8835" (UID: "6fa01532-be1d-4542-bad2-3608980a8835"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.531726 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5592df9-94ea-4dd6-9f76-473b02a44897-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5592df9-94ea-4dd6-9f76-473b02a44897" (UID: "b5592df9-94ea-4dd6-9f76-473b02a44897"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.538143 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fa01532-be1d-4542-bad2-3608980a8835-kube-api-access-zvczq" (OuterVolumeSpecName: "kube-api-access-zvczq") pod "6fa01532-be1d-4542-bad2-3608980a8835" (UID: "6fa01532-be1d-4542-bad2-3608980a8835"). InnerVolumeSpecName "kube-api-access-zvczq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.539449 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5592df9-94ea-4dd6-9f76-473b02a44897-kube-api-access-24795" (OuterVolumeSpecName: "kube-api-access-24795") pod "b5592df9-94ea-4dd6-9f76-473b02a44897" (UID: "b5592df9-94ea-4dd6-9f76-473b02a44897"). InnerVolumeSpecName "kube-api-access-24795". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.620551 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mk75p" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.620535 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mk75p" event={"ID":"6fa01532-be1d-4542-bad2-3608980a8835","Type":"ContainerDied","Data":"638a5a72c7c213261445134dc581b98f1dbf478ea16dd73e2ab2f9845497b406"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.620657 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="638a5a72c7c213261445134dc581b98f1dbf478ea16dd73e2ab2f9845497b406" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.622067 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qbr2v" event={"ID":"b5592df9-94ea-4dd6-9f76-473b02a44897","Type":"ContainerDied","Data":"908aba6352453cf2813c6304c3364f57f748e082b0c5543416eded9c86f227af"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.622105 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="908aba6352453cf2813c6304c3364f57f748e082b0c5543416eded9c86f227af" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.622176 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qbr2v" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.628180 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" event={"ID":"bb9c5d54-6432-4f78-a4a1-9dc43ffff102","Type":"ContainerDied","Data":"2fdc7168af0d3de69a5ccb3941640bacf893e00ff017d536c9ed75695eb1b2f7"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.628209 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fdc7168af0d3de69a5ccb3941640bacf893e00ff017d536c9ed75695eb1b2f7" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.629510 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-99d5-account-create-update-j5z7r" event={"ID":"ed0bfa2e-a25b-4dce-86be-1401ac97f10a","Type":"ContainerDied","Data":"985244c746bc5955b1eb02e933b3dca30653ef7d644d5416e6e19bcd0438bda2"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.629527 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="985244c746bc5955b1eb02e933b3dca30653ef7d644d5416e6e19bcd0438bda2" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.630601 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0e75-account-create-update-x7wmg" event={"ID":"2fa26501-8eec-4719-9b4d-211dc1e24ec9","Type":"ContainerDied","Data":"2f8674b94f10deef94971203c1ccfd0868cd2be01ba08373f53c5613068cf91f"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.630637 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f8674b94f10deef94971203c1ccfd0868cd2be01ba08373f53c5613068cf91f" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.631567 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvczq\" (UniqueName: \"kubernetes.io/projected/6fa01532-be1d-4542-bad2-3608980a8835-kube-api-access-zvczq\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.631585 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24795\" (UniqueName: \"kubernetes.io/projected/b5592df9-94ea-4dd6-9f76-473b02a44897-kube-api-access-24795\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.631595 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5592df9-94ea-4dd6-9f76-473b02a44897-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.631604 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fa01532-be1d-4542-bad2-3608980a8835-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.632049 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-47f3-account-create-update-sk9vk" event={"ID":"d6dcffe4-90e2-4334-88c9-60b336d87402","Type":"ContainerDied","Data":"b3bba2c1493986393ddd02b557cf6e9e08a325ec8515203571982d04cddd3536"} Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.632069 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3bba2c1493986393ddd02b557cf6e9e08a325ec8515203571982d04cddd3536" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.655870 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.663061 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.678526 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.696267 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.733258 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmlxs\" (UniqueName: \"kubernetes.io/projected/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-kube-api-access-kmlxs\") pod \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\" (UID: \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.733488 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-operator-scripts\") pod \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\" (UID: \"ed0bfa2e-a25b-4dce-86be-1401ac97f10a\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.734354 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed0bfa2e-a25b-4dce-86be-1401ac97f10a" (UID: "ed0bfa2e-a25b-4dce-86be-1401ac97f10a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.738446 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-kube-api-access-kmlxs" (OuterVolumeSpecName: "kube-api-access-kmlxs") pod "ed0bfa2e-a25b-4dce-86be-1401ac97f10a" (UID: "ed0bfa2e-a25b-4dce-86be-1401ac97f10a"). InnerVolumeSpecName "kube-api-access-kmlxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.834601 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89rtf\" (UniqueName: \"kubernetes.io/projected/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-kube-api-access-89rtf\") pod \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\" (UID: \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.835226 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4lq5\" (UniqueName: \"kubernetes.io/projected/2fa26501-8eec-4719-9b4d-211dc1e24ec9-kube-api-access-f4lq5\") pod \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\" (UID: \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.835279 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnzrx\" (UniqueName: \"kubernetes.io/projected/d6dcffe4-90e2-4334-88c9-60b336d87402-kube-api-access-bnzrx\") pod \"d6dcffe4-90e2-4334-88c9-60b336d87402\" (UID: \"d6dcffe4-90e2-4334-88c9-60b336d87402\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.835303 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa26501-8eec-4719-9b4d-211dc1e24ec9-operator-scripts\") pod \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\" (UID: \"2fa26501-8eec-4719-9b4d-211dc1e24ec9\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.835325 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6dcffe4-90e2-4334-88c9-60b336d87402-operator-scripts\") pod \"d6dcffe4-90e2-4334-88c9-60b336d87402\" (UID: \"d6dcffe4-90e2-4334-88c9-60b336d87402\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.835370 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-operator-scripts\") pod \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\" (UID: \"bb9c5d54-6432-4f78-a4a1-9dc43ffff102\") " Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.835864 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.835884 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmlxs\" (UniqueName: \"kubernetes.io/projected/ed0bfa2e-a25b-4dce-86be-1401ac97f10a-kube-api-access-kmlxs\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.836035 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6dcffe4-90e2-4334-88c9-60b336d87402-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6dcffe4-90e2-4334-88c9-60b336d87402" (UID: "d6dcffe4-90e2-4334-88c9-60b336d87402"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.836159 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fa26501-8eec-4719-9b4d-211dc1e24ec9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2fa26501-8eec-4719-9b4d-211dc1e24ec9" (UID: "2fa26501-8eec-4719-9b4d-211dc1e24ec9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.836181 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bb9c5d54-6432-4f78-a4a1-9dc43ffff102" (UID: "bb9c5d54-6432-4f78-a4a1-9dc43ffff102"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.838551 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-kube-api-access-89rtf" (OuterVolumeSpecName: "kube-api-access-89rtf") pod "bb9c5d54-6432-4f78-a4a1-9dc43ffff102" (UID: "bb9c5d54-6432-4f78-a4a1-9dc43ffff102"). InnerVolumeSpecName "kube-api-access-89rtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.838813 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6dcffe4-90e2-4334-88c9-60b336d87402-kube-api-access-bnzrx" (OuterVolumeSpecName: "kube-api-access-bnzrx") pod "d6dcffe4-90e2-4334-88c9-60b336d87402" (UID: "d6dcffe4-90e2-4334-88c9-60b336d87402"). InnerVolumeSpecName "kube-api-access-bnzrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.840224 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fa26501-8eec-4719-9b4d-211dc1e24ec9-kube-api-access-f4lq5" (OuterVolumeSpecName: "kube-api-access-f4lq5") pod "2fa26501-8eec-4719-9b4d-211dc1e24ec9" (UID: "2fa26501-8eec-4719-9b4d-211dc1e24ec9"). InnerVolumeSpecName "kube-api-access-f4lq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.938324 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89rtf\" (UniqueName: \"kubernetes.io/projected/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-kube-api-access-89rtf\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.938363 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnzrx\" (UniqueName: \"kubernetes.io/projected/d6dcffe4-90e2-4334-88c9-60b336d87402-kube-api-access-bnzrx\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.938377 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4lq5\" (UniqueName: \"kubernetes.io/projected/2fa26501-8eec-4719-9b4d-211dc1e24ec9-kube-api-access-f4lq5\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.938390 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fa26501-8eec-4719-9b4d-211dc1e24ec9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.938403 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6dcffe4-90e2-4334-88c9-60b336d87402-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:06 crc kubenswrapper[4848]: I1204 14:11:06.938416 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb9c5d54-6432-4f78-a4a1-9dc43ffff102-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.012033 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.053230 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.141495 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-dispersionconf\") pod \"ff826431-be61-4dea-82e7-b6849d838e52\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.141700 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlfk8\" (UniqueName: \"kubernetes.io/projected/7e78126d-0db0-4fee-a1b6-c23818909b5a-kube-api-access-qlfk8\") pod \"7e78126d-0db0-4fee-a1b6-c23818909b5a\" (UID: \"7e78126d-0db0-4fee-a1b6-c23818909b5a\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.141820 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-combined-ca-bundle\") pod \"ff826431-be61-4dea-82e7-b6849d838e52\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.141864 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-ring-data-devices\") pod \"ff826431-be61-4dea-82e7-b6849d838e52\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.141999 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-swiftconf\") pod \"ff826431-be61-4dea-82e7-b6849d838e52\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.142131 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-scripts\") pod \"ff826431-be61-4dea-82e7-b6849d838e52\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.142173 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e78126d-0db0-4fee-a1b6-c23818909b5a-operator-scripts\") pod \"7e78126d-0db0-4fee-a1b6-c23818909b5a\" (UID: \"7e78126d-0db0-4fee-a1b6-c23818909b5a\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.142271 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6zqx\" (UniqueName: \"kubernetes.io/projected/ff826431-be61-4dea-82e7-b6849d838e52-kube-api-access-f6zqx\") pod \"ff826431-be61-4dea-82e7-b6849d838e52\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.142585 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ff826431-be61-4dea-82e7-b6849d838e52" (UID: "ff826431-be61-4dea-82e7-b6849d838e52"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.142618 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e78126d-0db0-4fee-a1b6-c23818909b5a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7e78126d-0db0-4fee-a1b6-c23818909b5a" (UID: "7e78126d-0db0-4fee-a1b6-c23818909b5a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.142770 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ff826431-be61-4dea-82e7-b6849d838e52-etc-swift\") pod \"ff826431-be61-4dea-82e7-b6849d838e52\" (UID: \"ff826431-be61-4dea-82e7-b6849d838e52\") " Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.143697 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff826431-be61-4dea-82e7-b6849d838e52-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ff826431-be61-4dea-82e7-b6849d838e52" (UID: "ff826431-be61-4dea-82e7-b6849d838e52"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.144330 4848 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.144983 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7e78126d-0db0-4fee-a1b6-c23818909b5a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.145421 4848 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ff826431-be61-4dea-82e7-b6849d838e52-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.146242 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff826431-be61-4dea-82e7-b6849d838e52-kube-api-access-f6zqx" (OuterVolumeSpecName: "kube-api-access-f6zqx") pod "ff826431-be61-4dea-82e7-b6849d838e52" (UID: "ff826431-be61-4dea-82e7-b6849d838e52"). InnerVolumeSpecName "kube-api-access-f6zqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.146732 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e78126d-0db0-4fee-a1b6-c23818909b5a-kube-api-access-qlfk8" (OuterVolumeSpecName: "kube-api-access-qlfk8") pod "7e78126d-0db0-4fee-a1b6-c23818909b5a" (UID: "7e78126d-0db0-4fee-a1b6-c23818909b5a"). InnerVolumeSpecName "kube-api-access-qlfk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.152713 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ff826431-be61-4dea-82e7-b6849d838e52" (UID: "ff826431-be61-4dea-82e7-b6849d838e52"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.168543 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ff826431-be61-4dea-82e7-b6849d838e52" (UID: "ff826431-be61-4dea-82e7-b6849d838e52"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.169211 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff826431-be61-4dea-82e7-b6849d838e52" (UID: "ff826431-be61-4dea-82e7-b6849d838e52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.193598 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-scripts" (OuterVolumeSpecName: "scripts") pod "ff826431-be61-4dea-82e7-b6849d838e52" (UID: "ff826431-be61-4dea-82e7-b6849d838e52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.248591 4848 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.248634 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlfk8\" (UniqueName: \"kubernetes.io/projected/7e78126d-0db0-4fee-a1b6-c23818909b5a-kube-api-access-qlfk8\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.248666 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.248678 4848 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ff826431-be61-4dea-82e7-b6849d838e52-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.248687 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff826431-be61-4dea-82e7-b6849d838e52-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.248695 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6zqx\" (UniqueName: \"kubernetes.io/projected/ff826431-be61-4dea-82e7-b6849d838e52-kube-api-access-f6zqx\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.623235 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.649004 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" event={"ID":"7e78126d-0db0-4fee-a1b6-c23818909b5a","Type":"ContainerDied","Data":"b2f442bafab8c1b12a4224b531b9c9c36961ac391fdbd4e75ee9d25c23be6922"} Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.649054 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2f442bafab8c1b12a4224b531b9c9c36961ac391fdbd4e75ee9d25c23be6922" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.649113 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-fd6f-account-create-update-wwnqm" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.655225 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0e75-account-create-update-x7wmg" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.656160 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gplgt" event={"ID":"ff826431-be61-4dea-82e7-b6849d838e52","Type":"ContainerDied","Data":"913f947594ec9e9c4da5dac9215dd5c77588bb5b2cb747693c4db75904a84b64"} Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.656403 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="913f947594ec9e9c4da5dac9215dd5c77588bb5b2cb747693c4db75904a84b64" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.656331 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-47f3-account-create-update-sk9vk" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.656444 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-99d5-account-create-update-j5z7r" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.656276 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-cgjg9" Dec 04 14:11:07 crc kubenswrapper[4848]: I1204 14:11:07.656531 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gplgt" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.286997 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.500848 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.561030 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.814073 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-h29cq"] Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815080 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed0bfa2e-a25b-4dce-86be-1401ac97f10a" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815102 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed0bfa2e-a25b-4dce-86be-1401ac97f10a" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815111 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e78126d-0db0-4fee-a1b6-c23818909b5a" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815126 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e78126d-0db0-4fee-a1b6-c23818909b5a" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815148 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6dcffe4-90e2-4334-88c9-60b336d87402" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815154 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6dcffe4-90e2-4334-88c9-60b336d87402" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815168 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5592df9-94ea-4dd6-9f76-473b02a44897" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815175 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5592df9-94ea-4dd6-9f76-473b02a44897" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815185 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff826431-be61-4dea-82e7-b6849d838e52" containerName="swift-ring-rebalance" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815191 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff826431-be61-4dea-82e7-b6849d838e52" containerName="swift-ring-rebalance" Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815202 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fa26501-8eec-4719-9b4d-211dc1e24ec9" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815209 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fa26501-8eec-4719-9b4d-211dc1e24ec9" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815218 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa01532-be1d-4542-bad2-3608980a8835" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815225 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa01532-be1d-4542-bad2-3608980a8835" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815237 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb9c5d54-6432-4f78-a4a1-9dc43ffff102" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815242 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb9c5d54-6432-4f78-a4a1-9dc43ffff102" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: E1204 14:11:08.815251 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a530fe44-72a5-487a-8b28-fae47fa307c5" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815265 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a530fe44-72a5-487a-8b28-fae47fa307c5" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815447 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5592df9-94ea-4dd6-9f76-473b02a44897" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815459 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6dcffe4-90e2-4334-88c9-60b336d87402" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815473 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed0bfa2e-a25b-4dce-86be-1401ac97f10a" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815482 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fa26501-8eec-4719-9b4d-211dc1e24ec9" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815489 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fa01532-be1d-4542-bad2-3608980a8835" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815499 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a530fe44-72a5-487a-8b28-fae47fa307c5" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815507 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb9c5d54-6432-4f78-a4a1-9dc43ffff102" containerName="mariadb-database-create" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815518 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff826431-be61-4dea-82e7-b6849d838e52" containerName="swift-ring-rebalance" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.815528 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e78126d-0db0-4fee-a1b6-c23818909b5a" containerName="mariadb-account-create-update" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.828189 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-h29cq"] Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.828306 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.833233 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.899430 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpdmq\" (UniqueName: \"kubernetes.io/projected/a3a03294-8b5e-4985-9fcc-d32085691310-kube-api-access-zpdmq\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.899514 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-config\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.899567 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:08 crc kubenswrapper[4848]: I1204 14:11:08.899594 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.000174 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-b9f5p"] Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.001092 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpdmq\" (UniqueName: \"kubernetes.io/projected/a3a03294-8b5e-4985-9fcc-d32085691310-kube-api-access-zpdmq\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.001162 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-config\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.001205 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.001230 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.001804 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.002208 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.002590 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.002825 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-config\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.003932 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.015776 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-b9f5p"] Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.049725 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpdmq\" (UniqueName: \"kubernetes.io/projected/a3a03294-8b5e-4985-9fcc-d32085691310-kube-api-access-zpdmq\") pod \"dnsmasq-dns-6c89d5d749-h29cq\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.103131 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf1c526-2dfe-4c81-9846-c122113ebaf1-combined-ca-bundle\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.103228 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf1c526-2dfe-4c81-9846-c122113ebaf1-config\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.103273 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/caf1c526-2dfe-4c81-9846-c122113ebaf1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.103341 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/caf1c526-2dfe-4c81-9846-c122113ebaf1-ovs-rundir\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.103368 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzq55\" (UniqueName: \"kubernetes.io/projected/caf1c526-2dfe-4c81-9846-c122113ebaf1-kube-api-access-jzq55\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.103386 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/caf1c526-2dfe-4c81-9846-c122113ebaf1-ovn-rundir\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.148400 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.184466 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-h29cq"] Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.205196 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/caf1c526-2dfe-4c81-9846-c122113ebaf1-ovs-rundir\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.205250 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzq55\" (UniqueName: \"kubernetes.io/projected/caf1c526-2dfe-4c81-9846-c122113ebaf1-kube-api-access-jzq55\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.205273 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/caf1c526-2dfe-4c81-9846-c122113ebaf1-ovn-rundir\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.205322 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf1c526-2dfe-4c81-9846-c122113ebaf1-combined-ca-bundle\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.205394 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf1c526-2dfe-4c81-9846-c122113ebaf1-config\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.205448 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/caf1c526-2dfe-4c81-9846-c122113ebaf1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.206248 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/caf1c526-2dfe-4c81-9846-c122113ebaf1-ovn-rundir\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.206323 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/caf1c526-2dfe-4c81-9846-c122113ebaf1-ovs-rundir\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.207135 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf1c526-2dfe-4c81-9846-c122113ebaf1-config\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.213295 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/caf1c526-2dfe-4c81-9846-c122113ebaf1-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.215195 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf1c526-2dfe-4c81-9846-c122113ebaf1-combined-ca-bundle\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.233044 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzq55\" (UniqueName: \"kubernetes.io/projected/caf1c526-2dfe-4c81-9846-c122113ebaf1-kube-api-access-jzq55\") pod \"ovn-controller-metrics-b9f5p\" (UID: \"caf1c526-2dfe-4c81-9846-c122113ebaf1\") " pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.233304 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-x9k96"] Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.239125 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.241364 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.248875 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-x9k96"] Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.306866 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-config\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.306946 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.307008 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54grj\" (UniqueName: \"kubernetes.io/projected/c8675f3d-03ac-46e5-be9e-84991aa3e80a-kube-api-access-54grj\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.307044 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.307162 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-dns-svc\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.328803 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-b9f5p" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.369214 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.408527 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-dns-svc\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.408965 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-config\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.409037 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.409090 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54grj\" (UniqueName: \"kubernetes.io/projected/c8675f3d-03ac-46e5-be9e-84991aa3e80a-kube-api-access-54grj\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.409417 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.409603 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-dns-svc\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.409833 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-config\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.410040 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.411564 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.419490 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.436033 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54grj\" (UniqueName: \"kubernetes.io/projected/c8675f3d-03ac-46e5-be9e-84991aa3e80a-kube-api-access-54grj\") pod \"dnsmasq-dns-698758b865-x9k96\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.590177 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.597309 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.599142 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.599906 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-dk99n" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.599971 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.609409 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.623106 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.637918 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.714783 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c7da8d50-9ec2-4b1d-b373-e65541b02306-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.714837 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7da8d50-9ec2-4b1d-b373-e65541b02306-config\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.715113 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.715185 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.715315 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.715440 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7da8d50-9ec2-4b1d-b373-e65541b02306-scripts\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.715595 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llskw\" (UniqueName: \"kubernetes.io/projected/c7da8d50-9ec2-4b1d-b373-e65541b02306-kube-api-access-llskw\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.730480 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-h29cq"] Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.816977 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c7da8d50-9ec2-4b1d-b373-e65541b02306-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.817028 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7da8d50-9ec2-4b1d-b373-e65541b02306-config\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.817226 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.817850 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.817899 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7da8d50-9ec2-4b1d-b373-e65541b02306-config\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.817920 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.818016 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7da8d50-9ec2-4b1d-b373-e65541b02306-scripts\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.818079 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llskw\" (UniqueName: \"kubernetes.io/projected/c7da8d50-9ec2-4b1d-b373-e65541b02306-kube-api-access-llskw\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.818730 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c7da8d50-9ec2-4b1d-b373-e65541b02306-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.819472 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7da8d50-9ec2-4b1d-b373-e65541b02306-scripts\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.820693 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.821309 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.830031 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7da8d50-9ec2-4b1d-b373-e65541b02306-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.842015 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llskw\" (UniqueName: \"kubernetes.io/projected/c7da8d50-9ec2-4b1d-b373-e65541b02306-kube-api-access-llskw\") pod \"ovn-northd-0\" (UID: \"c7da8d50-9ec2-4b1d-b373-e65541b02306\") " pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.921269 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 14:11:09 crc kubenswrapper[4848]: I1204 14:11:09.932116 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-b9f5p"] Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.109660 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-x9k96"] Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.421860 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 14:11:10 crc kubenswrapper[4848]: W1204 14:11:10.426788 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7da8d50_9ec2_4b1d_b373_e65541b02306.slice/crio-b6ee101c0df101fef7e3044393f370da57cfcf2d1ba5752880eeeb871f5901bd WatchSource:0}: Error finding container b6ee101c0df101fef7e3044393f370da57cfcf2d1ba5752880eeeb871f5901bd: Status 404 returned error can't find the container with id b6ee101c0df101fef7e3044393f370da57cfcf2d1ba5752880eeeb871f5901bd Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.714975 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-b9f5p" event={"ID":"caf1c526-2dfe-4c81-9846-c122113ebaf1","Type":"ContainerStarted","Data":"f1d5b664ea170b3aaca3ed399cc2249bfd21fb93976a426ec6af41a99fc0f6a5"} Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.715348 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-b9f5p" event={"ID":"caf1c526-2dfe-4c81-9846-c122113ebaf1","Type":"ContainerStarted","Data":"552208ab2454e62ebc3deb6489ce278ae32ba0cc0b2aad4865a28b5f85433874"} Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.717530 4848 generic.go:334] "Generic (PLEG): container finished" podID="a3a03294-8b5e-4985-9fcc-d32085691310" containerID="13f30fe56fa9e6e907d18925834a21522d1a6e057bc88aec44b0c42eb58c3493" exitCode=0 Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.717596 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" event={"ID":"a3a03294-8b5e-4985-9fcc-d32085691310","Type":"ContainerDied","Data":"13f30fe56fa9e6e907d18925834a21522d1a6e057bc88aec44b0c42eb58c3493"} Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.717622 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" event={"ID":"a3a03294-8b5e-4985-9fcc-d32085691310","Type":"ContainerStarted","Data":"540912ef815731fa1e8b169dd210ec384ad1a6f40ac7f04b322ea835a8ff854b"} Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.722173 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-x9k96" event={"ID":"c8675f3d-03ac-46e5-be9e-84991aa3e80a","Type":"ContainerDied","Data":"b79aef1f5f288e7a71d74b5d217ffb3f69e3654849b2a7e05b2a2c3f8491200e"} Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.722028 4848 generic.go:334] "Generic (PLEG): container finished" podID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerID="b79aef1f5f288e7a71d74b5d217ffb3f69e3654849b2a7e05b2a2c3f8491200e" exitCode=0 Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.723190 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-x9k96" event={"ID":"c8675f3d-03ac-46e5-be9e-84991aa3e80a","Type":"ContainerStarted","Data":"7b7f0648a88dcaf1406f9f211e6ffacc6ee471c71d70b23ba7108b2e92bc6cf8"} Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.728479 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c7da8d50-9ec2-4b1d-b373-e65541b02306","Type":"ContainerStarted","Data":"b6ee101c0df101fef7e3044393f370da57cfcf2d1ba5752880eeeb871f5901bd"} Dec 04 14:11:10 crc kubenswrapper[4848]: I1204 14:11:10.759879 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-b9f5p" podStartSLOduration=2.759858623 podStartE2EDuration="2.759858623s" podCreationTimestamp="2025-12-04 14:11:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:10.736861846 +0000 UTC m=+1374.679358384" watchObservedRunningTime="2025-12-04 14:11:10.759858623 +0000 UTC m=+1374.702355171" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.213257 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.269530 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-ovsdbserver-sb\") pod \"a3a03294-8b5e-4985-9fcc-d32085691310\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.270519 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-dns-svc\") pod \"a3a03294-8b5e-4985-9fcc-d32085691310\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.270569 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpdmq\" (UniqueName: \"kubernetes.io/projected/a3a03294-8b5e-4985-9fcc-d32085691310-kube-api-access-zpdmq\") pod \"a3a03294-8b5e-4985-9fcc-d32085691310\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.270593 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-config\") pod \"a3a03294-8b5e-4985-9fcc-d32085691310\" (UID: \"a3a03294-8b5e-4985-9fcc-d32085691310\") " Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.280239 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a03294-8b5e-4985-9fcc-d32085691310-kube-api-access-zpdmq" (OuterVolumeSpecName: "kube-api-access-zpdmq") pod "a3a03294-8b5e-4985-9fcc-d32085691310" (UID: "a3a03294-8b5e-4985-9fcc-d32085691310"). InnerVolumeSpecName "kube-api-access-zpdmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.311564 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-config" (OuterVolumeSpecName: "config") pod "a3a03294-8b5e-4985-9fcc-d32085691310" (UID: "a3a03294-8b5e-4985-9fcc-d32085691310"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.319253 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a3a03294-8b5e-4985-9fcc-d32085691310" (UID: "a3a03294-8b5e-4985-9fcc-d32085691310"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.321609 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a3a03294-8b5e-4985-9fcc-d32085691310" (UID: "a3a03294-8b5e-4985-9fcc-d32085691310"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.374109 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.374151 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpdmq\" (UniqueName: \"kubernetes.io/projected/a3a03294-8b5e-4985-9fcc-d32085691310-kube-api-access-zpdmq\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.374168 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.374182 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a03294-8b5e-4985-9fcc-d32085691310-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.571125 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-r7w8t"] Dec 04 14:11:11 crc kubenswrapper[4848]: E1204 14:11:11.571625 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a03294-8b5e-4985-9fcc-d32085691310" containerName="init" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.571642 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a03294-8b5e-4985-9fcc-d32085691310" containerName="init" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.571926 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a03294-8b5e-4985-9fcc-d32085691310" containerName="init" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.572771 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.574479 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rzzqm" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.576531 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.580991 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-r7w8t"] Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.689898 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw2gh\" (UniqueName: \"kubernetes.io/projected/9925991d-d748-4a39-96a0-3bf2e391cad7-kube-api-access-qw2gh\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.690355 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-combined-ca-bundle\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.690378 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-db-sync-config-data\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.690431 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-config-data\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.738221 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" event={"ID":"a3a03294-8b5e-4985-9fcc-d32085691310","Type":"ContainerDied","Data":"540912ef815731fa1e8b169dd210ec384ad1a6f40ac7f04b322ea835a8ff854b"} Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.738278 4848 scope.go:117] "RemoveContainer" containerID="13f30fe56fa9e6e907d18925834a21522d1a6e057bc88aec44b0c42eb58c3493" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.738499 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-h29cq" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.741072 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-x9k96" event={"ID":"c8675f3d-03ac-46e5-be9e-84991aa3e80a","Type":"ContainerStarted","Data":"1e568a7ab9c8b37d95568cc59c1b6e0a0a7d7328b565bd47cb8c987ae3780f0d"} Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.741201 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.770391 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-x9k96" podStartSLOduration=2.770377093 podStartE2EDuration="2.770377093s" podCreationTimestamp="2025-12-04 14:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:11.770180418 +0000 UTC m=+1375.712676946" watchObservedRunningTime="2025-12-04 14:11:11.770377093 +0000 UTC m=+1375.712873621" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.792387 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-combined-ca-bundle\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.792677 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-db-sync-config-data\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.792816 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-config-data\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.793028 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw2gh\" (UniqueName: \"kubernetes.io/projected/9925991d-d748-4a39-96a0-3bf2e391cad7-kube-api-access-qw2gh\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.800776 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-config-data\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.803901 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-combined-ca-bundle\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.805200 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-db-sync-config-data\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.813600 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw2gh\" (UniqueName: \"kubernetes.io/projected/9925991d-d748-4a39-96a0-3bf2e391cad7-kube-api-access-qw2gh\") pod \"glance-db-sync-r7w8t\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.860243 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-h29cq"] Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.882483 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-h29cq"] Dec 04 14:11:11 crc kubenswrapper[4848]: I1204 14:11:11.894860 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:12 crc kubenswrapper[4848]: I1204 14:11:12.404387 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a03294-8b5e-4985-9fcc-d32085691310" path="/var/lib/kubelet/pods/a3a03294-8b5e-4985-9fcc-d32085691310/volumes" Dec 04 14:11:12 crc kubenswrapper[4848]: I1204 14:11:12.405572 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-r7w8t"] Dec 04 14:11:12 crc kubenswrapper[4848]: I1204 14:11:12.752381 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r7w8t" event={"ID":"9925991d-d748-4a39-96a0-3bf2e391cad7","Type":"ContainerStarted","Data":"50a37c81c8ab5d5bedaa655a73650b51d6e7b4d285da531ee9d8cf62b0ffae67"} Dec 04 14:11:12 crc kubenswrapper[4848]: I1204 14:11:12.756001 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c7da8d50-9ec2-4b1d-b373-e65541b02306","Type":"ContainerStarted","Data":"4c5b3f51fb0d36eee34a7d1568e3e4d7a988b7ff54344291dd70800a3dc3df6f"} Dec 04 14:11:12 crc kubenswrapper[4848]: I1204 14:11:12.756051 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c7da8d50-9ec2-4b1d-b373-e65541b02306","Type":"ContainerStarted","Data":"a368daadb3b1259918905309a9ebf5df0460de07f50d4ab8d0d6b76407a799b5"} Dec 04 14:11:12 crc kubenswrapper[4848]: I1204 14:11:12.782796 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.666181247 podStartE2EDuration="3.782775788s" podCreationTimestamp="2025-12-04 14:11:09 +0000 UTC" firstStartedPulling="2025-12-04 14:11:10.428807895 +0000 UTC m=+1374.371304423" lastFinishedPulling="2025-12-04 14:11:11.545402436 +0000 UTC m=+1375.487898964" observedRunningTime="2025-12-04 14:11:12.775657845 +0000 UTC m=+1376.718154393" watchObservedRunningTime="2025-12-04 14:11:12.782775788 +0000 UTC m=+1376.725272316" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.138569 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4"] Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.139862 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.151818 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4"] Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.219767 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms6j4\" (UniqueName: \"kubernetes.io/projected/48d945e4-9311-4d09-a75a-d6d1b5e59dff-kube-api-access-ms6j4\") pod \"mysqld-exporter-openstack-cell1-db-create-bxkz4\" (UID: \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.219933 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48d945e4-9311-4d09-a75a-d6d1b5e59dff-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-bxkz4\" (UID: \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.322286 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms6j4\" (UniqueName: \"kubernetes.io/projected/48d945e4-9311-4d09-a75a-d6d1b5e59dff-kube-api-access-ms6j4\") pod \"mysqld-exporter-openstack-cell1-db-create-bxkz4\" (UID: \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.322506 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48d945e4-9311-4d09-a75a-d6d1b5e59dff-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-bxkz4\" (UID: \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.323238 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48d945e4-9311-4d09-a75a-d6d1b5e59dff-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-bxkz4\" (UID: \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.351921 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms6j4\" (UniqueName: \"kubernetes.io/projected/48d945e4-9311-4d09-a75a-d6d1b5e59dff-kube-api-access-ms6j4\") pod \"mysqld-exporter-openstack-cell1-db-create-bxkz4\" (UID: \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.361005 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-7f5b-account-create-update-9j2qq"] Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.362414 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.366872 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.372432 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-7f5b-account-create-update-9j2qq"] Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.424763 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvdkd\" (UniqueName: \"kubernetes.io/projected/e197c4da-c849-4c84-becb-dde8501eda45-kube-api-access-lvdkd\") pod \"mysqld-exporter-7f5b-account-create-update-9j2qq\" (UID: \"e197c4da-c849-4c84-becb-dde8501eda45\") " pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.425628 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e197c4da-c849-4c84-becb-dde8501eda45-operator-scripts\") pod \"mysqld-exporter-7f5b-account-create-update-9j2qq\" (UID: \"e197c4da-c849-4c84-becb-dde8501eda45\") " pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.482229 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.528805 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e197c4da-c849-4c84-becb-dde8501eda45-operator-scripts\") pod \"mysqld-exporter-7f5b-account-create-update-9j2qq\" (UID: \"e197c4da-c849-4c84-becb-dde8501eda45\") " pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.529046 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvdkd\" (UniqueName: \"kubernetes.io/projected/e197c4da-c849-4c84-becb-dde8501eda45-kube-api-access-lvdkd\") pod \"mysqld-exporter-7f5b-account-create-update-9j2qq\" (UID: \"e197c4da-c849-4c84-becb-dde8501eda45\") " pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.530437 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e197c4da-c849-4c84-becb-dde8501eda45-operator-scripts\") pod \"mysqld-exporter-7f5b-account-create-update-9j2qq\" (UID: \"e197c4da-c849-4c84-becb-dde8501eda45\") " pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.549807 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvdkd\" (UniqueName: \"kubernetes.io/projected/e197c4da-c849-4c84-becb-dde8501eda45-kube-api-access-lvdkd\") pod \"mysqld-exporter-7f5b-account-create-update-9j2qq\" (UID: \"e197c4da-c849-4c84-becb-dde8501eda45\") " pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.716767 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.778886 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 04 14:11:13 crc kubenswrapper[4848]: I1204 14:11:13.924059 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4"] Dec 04 14:11:13 crc kubenswrapper[4848]: W1204 14:11:13.935544 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48d945e4_9311_4d09_a75a_d6d1b5e59dff.slice/crio-5d27ab83fe7fd84d879ae821adada6d96cd20527e8685ff0c84b6db1f164c4e9 WatchSource:0}: Error finding container 5d27ab83fe7fd84d879ae821adada6d96cd20527e8685ff0c84b6db1f164c4e9: Status 404 returned error can't find the container with id 5d27ab83fe7fd84d879ae821adada6d96cd20527e8685ff0c84b6db1f164c4e9 Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.192302 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-7f5b-account-create-update-9j2qq"] Dec 04 14:11:14 crc kubenswrapper[4848]: W1204 14:11:14.198012 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode197c4da_c849_4c84_becb_dde8501eda45.slice/crio-a7ff3a4a685b9f891c51ebc557f2621f8d1463d8cbd3b14bd38d82851e8ee12e WatchSource:0}: Error finding container a7ff3a4a685b9f891c51ebc557f2621f8d1463d8cbd3b14bd38d82851e8ee12e: Status 404 returned error can't find the container with id a7ff3a4a685b9f891c51ebc557f2621f8d1463d8cbd3b14bd38d82851e8ee12e Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.313886 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.314144 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.472344 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.476610 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.811334 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" event={"ID":"48d945e4-9311-4d09-a75a-d6d1b5e59dff","Type":"ContainerStarted","Data":"9097b4a5e0afac0434bf3a3eb2919f2f21b0524f320a1334550aaee2873df5af"} Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.811391 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" event={"ID":"48d945e4-9311-4d09-a75a-d6d1b5e59dff","Type":"ContainerStarted","Data":"5d27ab83fe7fd84d879ae821adada6d96cd20527e8685ff0c84b6db1f164c4e9"} Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.812898 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" event={"ID":"e197c4da-c849-4c84-becb-dde8501eda45","Type":"ContainerStarted","Data":"72a027a9649a4c0f6fc55e4d5cb03312847100751315821ac11a94d579d2652d"} Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.812927 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" event={"ID":"e197c4da-c849-4c84-becb-dde8501eda45","Type":"ContainerStarted","Data":"a7ff3a4a685b9f891c51ebc557f2621f8d1463d8cbd3b14bd38d82851e8ee12e"} Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.814621 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.826106 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" podStartSLOduration=1.826089506 podStartE2EDuration="1.826089506s" podCreationTimestamp="2025-12-04 14:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:14.8245992 +0000 UTC m=+1378.767095728" watchObservedRunningTime="2025-12-04 14:11:14.826089506 +0000 UTC m=+1378.768586034" Dec 04 14:11:14 crc kubenswrapper[4848]: I1204 14:11:14.872045 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" podStartSLOduration=1.872026141 podStartE2EDuration="1.872026141s" podCreationTimestamp="2025-12-04 14:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:14.865423091 +0000 UTC m=+1378.807919619" watchObservedRunningTime="2025-12-04 14:11:14.872026141 +0000 UTC m=+1378.814522669" Dec 04 14:11:15 crc kubenswrapper[4848]: I1204 14:11:15.829646 4848 generic.go:334] "Generic (PLEG): container finished" podID="e197c4da-c849-4c84-becb-dde8501eda45" containerID="72a027a9649a4c0f6fc55e4d5cb03312847100751315821ac11a94d579d2652d" exitCode=0 Dec 04 14:11:15 crc kubenswrapper[4848]: I1204 14:11:15.829719 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" event={"ID":"e197c4da-c849-4c84-becb-dde8501eda45","Type":"ContainerDied","Data":"72a027a9649a4c0f6fc55e4d5cb03312847100751315821ac11a94d579d2652d"} Dec 04 14:11:15 crc kubenswrapper[4848]: I1204 14:11:15.838889 4848 generic.go:334] "Generic (PLEG): container finished" podID="48d945e4-9311-4d09-a75a-d6d1b5e59dff" containerID="9097b4a5e0afac0434bf3a3eb2919f2f21b0524f320a1334550aaee2873df5af" exitCode=0 Dec 04 14:11:15 crc kubenswrapper[4848]: I1204 14:11:15.839943 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" event={"ID":"48d945e4-9311-4d09-a75a-d6d1b5e59dff","Type":"ContainerDied","Data":"9097b4a5e0afac0434bf3a3eb2919f2f21b0524f320a1334550aaee2873df5af"} Dec 04 14:11:16 crc kubenswrapper[4848]: I1204 14:11:16.193846 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:11:16 crc kubenswrapper[4848]: I1204 14:11:16.201239 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0102871d-c950-46e7-af43-a2b3f1cbc39f-etc-swift\") pod \"swift-storage-0\" (UID: \"0102871d-c950-46e7-af43-a2b3f1cbc39f\") " pod="openstack/swift-storage-0" Dec 04 14:11:16 crc kubenswrapper[4848]: I1204 14:11:16.335345 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 14:11:16 crc kubenswrapper[4848]: I1204 14:11:16.977635 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 14:11:16 crc kubenswrapper[4848]: W1204 14:11:16.995899 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0102871d_c950_46e7_af43_a2b3f1cbc39f.slice/crio-f6b19655e28e724a17262fd4522550896fd3841e3445756b5a8b23236f3127c6 WatchSource:0}: Error finding container f6b19655e28e724a17262fd4522550896fd3841e3445756b5a8b23236f3127c6: Status 404 returned error can't find the container with id f6b19655e28e724a17262fd4522550896fd3841e3445756b5a8b23236f3127c6 Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.371333 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.377743 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.422619 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e197c4da-c849-4c84-becb-dde8501eda45-operator-scripts\") pod \"e197c4da-c849-4c84-becb-dde8501eda45\" (UID: \"e197c4da-c849-4c84-becb-dde8501eda45\") " Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.422796 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48d945e4-9311-4d09-a75a-d6d1b5e59dff-operator-scripts\") pod \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\" (UID: \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\") " Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.422830 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvdkd\" (UniqueName: \"kubernetes.io/projected/e197c4da-c849-4c84-becb-dde8501eda45-kube-api-access-lvdkd\") pod \"e197c4da-c849-4c84-becb-dde8501eda45\" (UID: \"e197c4da-c849-4c84-becb-dde8501eda45\") " Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.422868 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms6j4\" (UniqueName: \"kubernetes.io/projected/48d945e4-9311-4d09-a75a-d6d1b5e59dff-kube-api-access-ms6j4\") pod \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\" (UID: \"48d945e4-9311-4d09-a75a-d6d1b5e59dff\") " Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.423744 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48d945e4-9311-4d09-a75a-d6d1b5e59dff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "48d945e4-9311-4d09-a75a-d6d1b5e59dff" (UID: "48d945e4-9311-4d09-a75a-d6d1b5e59dff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.424261 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e197c4da-c849-4c84-becb-dde8501eda45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e197c4da-c849-4c84-becb-dde8501eda45" (UID: "e197c4da-c849-4c84-becb-dde8501eda45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.424812 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48d945e4-9311-4d09-a75a-d6d1b5e59dff-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.424829 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e197c4da-c849-4c84-becb-dde8501eda45-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.428571 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e197c4da-c849-4c84-becb-dde8501eda45-kube-api-access-lvdkd" (OuterVolumeSpecName: "kube-api-access-lvdkd") pod "e197c4da-c849-4c84-becb-dde8501eda45" (UID: "e197c4da-c849-4c84-becb-dde8501eda45"). InnerVolumeSpecName "kube-api-access-lvdkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.433925 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d945e4-9311-4d09-a75a-d6d1b5e59dff-kube-api-access-ms6j4" (OuterVolumeSpecName: "kube-api-access-ms6j4") pod "48d945e4-9311-4d09-a75a-d6d1b5e59dff" (UID: "48d945e4-9311-4d09-a75a-d6d1b5e59dff"). InnerVolumeSpecName "kube-api-access-ms6j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.536466 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvdkd\" (UniqueName: \"kubernetes.io/projected/e197c4da-c849-4c84-becb-dde8501eda45-kube-api-access-lvdkd\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.536516 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms6j4\" (UniqueName: \"kubernetes.io/projected/48d945e4-9311-4d09-a75a-d6d1b5e59dff-kube-api-access-ms6j4\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.805266 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.805721 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="prometheus" containerID="cri-o://57084ce6e036e4bb452a513c26295d75c4c3ba17caf688ccffdc8a807ef9505e" gracePeriod=600 Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.805811 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="thanos-sidecar" containerID="cri-o://14f013eceadc2ad0e7111f2fe485eefe5cb5a76b319302c45f850d5966f830ea" gracePeriod=600 Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.805811 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="config-reloader" containerID="cri-o://4a29902bc73018a50387e59b45b3c5a4eec4d42ef6f1c00e1528e361cd96a8c4" gracePeriod=600 Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.864213 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"f6b19655e28e724a17262fd4522550896fd3841e3445756b5a8b23236f3127c6"} Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.869315 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" event={"ID":"e197c4da-c849-4c84-becb-dde8501eda45","Type":"ContainerDied","Data":"a7ff3a4a685b9f891c51ebc557f2621f8d1463d8cbd3b14bd38d82851e8ee12e"} Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.869363 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7ff3a4a685b9f891c51ebc557f2621f8d1463d8cbd3b14bd38d82851e8ee12e" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.869382 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7f5b-account-create-update-9j2qq" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.871331 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" event={"ID":"48d945e4-9311-4d09-a75a-d6d1b5e59dff","Type":"ContainerDied","Data":"5d27ab83fe7fd84d879ae821adada6d96cd20527e8685ff0c84b6db1f164c4e9"} Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.871377 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d27ab83fe7fd84d879ae821adada6d96cd20527e8685ff0c84b6db1f164c4e9" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.871453 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4" Dec 04 14:11:17 crc kubenswrapper[4848]: I1204 14:11:17.940247 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.342784 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-61ea-account-create-update-mtrsm"] Dec 04 14:11:18 crc kubenswrapper[4848]: E1204 14:11:18.343439 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e197c4da-c849-4c84-becb-dde8501eda45" containerName="mariadb-account-create-update" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.343789 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="e197c4da-c849-4c84-becb-dde8501eda45" containerName="mariadb-account-create-update" Dec 04 14:11:18 crc kubenswrapper[4848]: E1204 14:11:18.343814 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d945e4-9311-4d09-a75a-d6d1b5e59dff" containerName="mariadb-database-create" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.343821 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d945e4-9311-4d09-a75a-d6d1b5e59dff" containerName="mariadb-database-create" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.344337 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d945e4-9311-4d09-a75a-d6d1b5e59dff" containerName="mariadb-database-create" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.344403 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="e197c4da-c849-4c84-becb-dde8501eda45" containerName="mariadb-account-create-update" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.346733 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.356295 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.378871 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-cxgrr"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.380642 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.392446 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-61ea-account-create-update-mtrsm"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.413252 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cxgrr"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.460878 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df74ef9c-2afd-4236-901e-82c48594f70f-operator-scripts\") pod \"cinder-61ea-account-create-update-mtrsm\" (UID: \"df74ef9c-2afd-4236-901e-82c48594f70f\") " pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.462683 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfhjr\" (UniqueName: \"kubernetes.io/projected/df74ef9c-2afd-4236-901e-82c48594f70f-kube-api-access-wfhjr\") pod \"cinder-61ea-account-create-update-mtrsm\" (UID: \"df74ef9c-2afd-4236-901e-82c48594f70f\") " pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.462917 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l44fp\" (UniqueName: \"kubernetes.io/projected/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-kube-api-access-l44fp\") pod \"cinder-db-create-cxgrr\" (UID: \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\") " pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.463661 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-operator-scripts\") pod \"cinder-db-create-cxgrr\" (UID: \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\") " pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.474487 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-2804-account-create-update-mrhhd"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.476210 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.477932 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.483170 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2804-account-create-update-mrhhd"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.563421 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-ldg2g"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.566190 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df74ef9c-2afd-4236-901e-82c48594f70f-operator-scripts\") pod \"cinder-61ea-account-create-update-mtrsm\" (UID: \"df74ef9c-2afd-4236-901e-82c48594f70f\") " pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.566252 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfhjr\" (UniqueName: \"kubernetes.io/projected/df74ef9c-2afd-4236-901e-82c48594f70f-kube-api-access-wfhjr\") pod \"cinder-61ea-account-create-update-mtrsm\" (UID: \"df74ef9c-2afd-4236-901e-82c48594f70f\") " pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.566291 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l44fp\" (UniqueName: \"kubernetes.io/projected/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-kube-api-access-l44fp\") pod \"cinder-db-create-cxgrr\" (UID: \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\") " pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.566370 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a756447-5ce8-4b33-9075-e3b6100d807c-operator-scripts\") pod \"barbican-2804-account-create-update-mrhhd\" (UID: \"6a756447-5ce8-4b33-9075-e3b6100d807c\") " pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.566420 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-operator-scripts\") pod \"cinder-db-create-cxgrr\" (UID: \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\") " pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.566468 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njm48\" (UniqueName: \"kubernetes.io/projected/6a756447-5ce8-4b33-9075-e3b6100d807c-kube-api-access-njm48\") pod \"barbican-2804-account-create-update-mrhhd\" (UID: \"6a756447-5ce8-4b33-9075-e3b6100d807c\") " pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.567249 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df74ef9c-2afd-4236-901e-82c48594f70f-operator-scripts\") pod \"cinder-61ea-account-create-update-mtrsm\" (UID: \"df74ef9c-2afd-4236-901e-82c48594f70f\") " pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.567740 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-operator-scripts\") pod \"cinder-db-create-cxgrr\" (UID: \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\") " pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.567841 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.587573 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-ldg2g"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.613674 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l44fp\" (UniqueName: \"kubernetes.io/projected/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-kube-api-access-l44fp\") pod \"cinder-db-create-cxgrr\" (UID: \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\") " pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.622721 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfhjr\" (UniqueName: \"kubernetes.io/projected/df74ef9c-2afd-4236-901e-82c48594f70f-kube-api-access-wfhjr\") pod \"cinder-61ea-account-create-update-mtrsm\" (UID: \"df74ef9c-2afd-4236-901e-82c48594f70f\") " pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.647749 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-8pzdg"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.649097 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.652184 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.652356 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4527c" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.656362 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.669316 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njm48\" (UniqueName: \"kubernetes.io/projected/6a756447-5ce8-4b33-9075-e3b6100d807c-kube-api-access-njm48\") pod \"barbican-2804-account-create-update-mrhhd\" (UID: \"6a756447-5ce8-4b33-9075-e3b6100d807c\") " pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.669474 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e722893-4a07-44e8-9e5a-3e04ff50dc47-operator-scripts\") pod \"barbican-db-create-ldg2g\" (UID: \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\") " pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.669499 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh27s\" (UniqueName: \"kubernetes.io/projected/4e722893-4a07-44e8-9e5a-3e04ff50dc47-kube-api-access-zh27s\") pod \"barbican-db-create-ldg2g\" (UID: \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\") " pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.669535 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a756447-5ce8-4b33-9075-e3b6100d807c-operator-scripts\") pod \"barbican-2804-account-create-update-mrhhd\" (UID: \"6a756447-5ce8-4b33-9075-e3b6100d807c\") " pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.670262 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a756447-5ce8-4b33-9075-e3b6100d807c-operator-scripts\") pod \"barbican-2804-account-create-update-mrhhd\" (UID: \"6a756447-5ce8-4b33-9075-e3b6100d807c\") " pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.671050 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.740891 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.741637 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-8pzdg"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.749309 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.771364 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-combined-ca-bundle\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.771417 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e722893-4a07-44e8-9e5a-3e04ff50dc47-operator-scripts\") pod \"barbican-db-create-ldg2g\" (UID: \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\") " pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.771439 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh27s\" (UniqueName: \"kubernetes.io/projected/4e722893-4a07-44e8-9e5a-3e04ff50dc47-kube-api-access-zh27s\") pod \"barbican-db-create-ldg2g\" (UID: \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\") " pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.771473 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tqnx\" (UniqueName: \"kubernetes.io/projected/cf2c3a14-5cfd-49a1-9337-fb0b61611669-kube-api-access-9tqnx\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.771507 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-config-data\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.772355 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e722893-4a07-44e8-9e5a-3e04ff50dc47-operator-scripts\") pod \"barbican-db-create-ldg2g\" (UID: \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\") " pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.873184 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-combined-ca-bundle\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.873255 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tqnx\" (UniqueName: \"kubernetes.io/projected/cf2c3a14-5cfd-49a1-9337-fb0b61611669-kube-api-access-9tqnx\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.873296 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-config-data\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.875749 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-dchv8"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.892318 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-dchv8" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.912472 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-dchv8"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.961789 4848 generic.go:334] "Generic (PLEG): container finished" podID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerID="14f013eceadc2ad0e7111f2fe485eefe5cb5a76b319302c45f850d5966f830ea" exitCode=0 Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.961818 4848 generic.go:334] "Generic (PLEG): container finished" podID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerID="4a29902bc73018a50387e59b45b3c5a4eec4d42ef6f1c00e1528e361cd96a8c4" exitCode=0 Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.961825 4848 generic.go:334] "Generic (PLEG): container finished" podID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerID="57084ce6e036e4bb452a513c26295d75c4c3ba17caf688ccffdc8a807ef9505e" exitCode=0 Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.961842 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerDied","Data":"14f013eceadc2ad0e7111f2fe485eefe5cb5a76b319302c45f850d5966f830ea"} Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.961865 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerDied","Data":"4a29902bc73018a50387e59b45b3c5a4eec4d42ef6f1c00e1528e361cd96a8c4"} Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.961874 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerDied","Data":"57084ce6e036e4bb452a513c26295d75c4c3ba17caf688ccffdc8a807ef9505e"} Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.972663 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njm48\" (UniqueName: \"kubernetes.io/projected/6a756447-5ce8-4b33-9075-e3b6100d807c-kube-api-access-njm48\") pod \"barbican-2804-account-create-update-mrhhd\" (UID: \"6a756447-5ce8-4b33-9075-e3b6100d807c\") " pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.973296 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh27s\" (UniqueName: \"kubernetes.io/projected/4e722893-4a07-44e8-9e5a-3e04ff50dc47-kube-api-access-zh27s\") pod \"barbican-db-create-ldg2g\" (UID: \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\") " pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.976857 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tqnx\" (UniqueName: \"kubernetes.io/projected/cf2c3a14-5cfd-49a1-9337-fb0b61611669-kube-api-access-9tqnx\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.977427 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-config-data\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.981642 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-operator-scripts\") pod \"heat-db-create-dchv8\" (UID: \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\") " pod="openstack/heat-db-create-dchv8" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.981741 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjjgb\" (UniqueName: \"kubernetes.io/projected/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-kube-api-access-tjjgb\") pod \"heat-db-create-dchv8\" (UID: \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\") " pod="openstack/heat-db-create-dchv8" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.982373 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-fd58-account-create-update-r8csb"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.983585 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.994770 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-fd58-account-create-update-r8csb"] Dec 04 14:11:18 crc kubenswrapper[4848]: I1204 14:11:18.996479 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-combined-ca-bundle\") pod \"keystone-db-sync-8pzdg\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.000478 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.001486 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.031791 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.072959 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.086543 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-operator-scripts\") pod \"heat-db-create-dchv8\" (UID: \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\") " pod="openstack/heat-db-create-dchv8" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.086617 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c742l\" (UniqueName: \"kubernetes.io/projected/41c0df85-a096-4e29-9dae-4e98007e43ff-kube-api-access-c742l\") pod \"heat-fd58-account-create-update-r8csb\" (UID: \"41c0df85-a096-4e29-9dae-4e98007e43ff\") " pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.086692 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41c0df85-a096-4e29-9dae-4e98007e43ff-operator-scripts\") pod \"heat-fd58-account-create-update-r8csb\" (UID: \"41c0df85-a096-4e29-9dae-4e98007e43ff\") " pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.086723 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjjgb\" (UniqueName: \"kubernetes.io/projected/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-kube-api-access-tjjgb\") pod \"heat-db-create-dchv8\" (UID: \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\") " pod="openstack/heat-db-create-dchv8" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.088128 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-operator-scripts\") pod \"heat-db-create-dchv8\" (UID: \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\") " pod="openstack/heat-db-create-dchv8" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.092509 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.098178 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.101518 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.127397 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.174880 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-s2z5f"] Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.198655 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjjgb\" (UniqueName: \"kubernetes.io/projected/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-kube-api-access-tjjgb\") pod \"heat-db-create-dchv8\" (UID: \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\") " pod="openstack/heat-db-create-dchv8" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.230971 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c742l\" (UniqueName: \"kubernetes.io/projected/41c0df85-a096-4e29-9dae-4e98007e43ff-kube-api-access-c742l\") pod \"heat-fd58-account-create-update-r8csb\" (UID: \"41c0df85-a096-4e29-9dae-4e98007e43ff\") " pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.231176 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41c0df85-a096-4e29-9dae-4e98007e43ff-operator-scripts\") pod \"heat-fd58-account-create-update-r8csb\" (UID: \"41c0df85-a096-4e29-9dae-4e98007e43ff\") " pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.237265 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.237424 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41c0df85-a096-4e29-9dae-4e98007e43ff-operator-scripts\") pod \"heat-fd58-account-create-update-r8csb\" (UID: \"41c0df85-a096-4e29-9dae-4e98007e43ff\") " pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.317442 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c742l\" (UniqueName: \"kubernetes.io/projected/41c0df85-a096-4e29-9dae-4e98007e43ff-kube-api-access-c742l\") pod \"heat-fd58-account-create-update-r8csb\" (UID: \"41c0df85-a096-4e29-9dae-4e98007e43ff\") " pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.323001 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-021e-account-create-update-g264j"] Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.324654 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.328906 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.331313 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s2z5f"] Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.332757 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsw2m\" (UniqueName: \"kubernetes.io/projected/cac2310b-905a-42df-92d8-611fa944c8f6-kube-api-access-tsw2m\") pod \"neutron-db-create-s2z5f\" (UID: \"cac2310b-905a-42df-92d8-611fa944c8f6\") " pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.332806 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/867de14d-11f1-4306-873d-cdc3f6f90216-operator-scripts\") pod \"neutron-021e-account-create-update-g264j\" (UID: \"867de14d-11f1-4306-873d-cdc3f6f90216\") " pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.332827 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c52ht\" (UniqueName: \"kubernetes.io/projected/867de14d-11f1-4306-873d-cdc3f6f90216-kube-api-access-c52ht\") pod \"neutron-021e-account-create-update-g264j\" (UID: \"867de14d-11f1-4306-873d-cdc3f6f90216\") " pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.332850 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-config-data\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.332875 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cac2310b-905a-42df-92d8-611fa944c8f6-operator-scripts\") pod \"neutron-db-create-s2z5f\" (UID: \"cac2310b-905a-42df-92d8-611fa944c8f6\") " pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.333178 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.333544 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvxwc\" (UniqueName: \"kubernetes.io/projected/d2af561d-8819-4175-89c0-4a65a661fdb6-kube-api-access-rvxwc\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.370333 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-021e-account-create-update-g264j"] Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.395585 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-dchv8" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.436019 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsw2m\" (UniqueName: \"kubernetes.io/projected/cac2310b-905a-42df-92d8-611fa944c8f6-kube-api-access-tsw2m\") pod \"neutron-db-create-s2z5f\" (UID: \"cac2310b-905a-42df-92d8-611fa944c8f6\") " pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.436080 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/867de14d-11f1-4306-873d-cdc3f6f90216-operator-scripts\") pod \"neutron-021e-account-create-update-g264j\" (UID: \"867de14d-11f1-4306-873d-cdc3f6f90216\") " pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.436119 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c52ht\" (UniqueName: \"kubernetes.io/projected/867de14d-11f1-4306-873d-cdc3f6f90216-kube-api-access-c52ht\") pod \"neutron-021e-account-create-update-g264j\" (UID: \"867de14d-11f1-4306-873d-cdc3f6f90216\") " pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.436149 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-config-data\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.436186 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cac2310b-905a-42df-92d8-611fa944c8f6-operator-scripts\") pod \"neutron-db-create-s2z5f\" (UID: \"cac2310b-905a-42df-92d8-611fa944c8f6\") " pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.436248 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.436366 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvxwc\" (UniqueName: \"kubernetes.io/projected/d2af561d-8819-4175-89c0-4a65a661fdb6-kube-api-access-rvxwc\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.437430 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/867de14d-11f1-4306-873d-cdc3f6f90216-operator-scripts\") pod \"neutron-021e-account-create-update-g264j\" (UID: \"867de14d-11f1-4306-873d-cdc3f6f90216\") " pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.449751 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cac2310b-905a-42df-92d8-611fa944c8f6-operator-scripts\") pod \"neutron-db-create-s2z5f\" (UID: \"cac2310b-905a-42df-92d8-611fa944c8f6\") " pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.456297 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-config-data\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.477168 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c52ht\" (UniqueName: \"kubernetes.io/projected/867de14d-11f1-4306-873d-cdc3f6f90216-kube-api-access-c52ht\") pod \"neutron-021e-account-create-update-g264j\" (UID: \"867de14d-11f1-4306-873d-cdc3f6f90216\") " pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.483036 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.484490 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.491077 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsw2m\" (UniqueName: \"kubernetes.io/projected/cac2310b-905a-42df-92d8-611fa944c8f6-kube-api-access-tsw2m\") pod \"neutron-db-create-s2z5f\" (UID: \"cac2310b-905a-42df-92d8-611fa944c8f6\") " pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.539934 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvxwc\" (UniqueName: \"kubernetes.io/projected/d2af561d-8819-4175-89c0-4a65a661fdb6-kube-api-access-rvxwc\") pod \"mysqld-exporter-0\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.652164 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.737062 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-2hkgw"] Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.737291 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" podUID="565e9c18-f099-49a4-9457-725b85e0cb7a" containerName="dnsmasq-dns" containerID="cri-o://def06344ecd9560b993c752d7c61cedd7d29be1747532632fd15618959d6afaf" gracePeriod=10 Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.908036 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.937330 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.963428 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:19 crc kubenswrapper[4848]: I1204 14:11:19.993226 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.042094 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1","Type":"ContainerDied","Data":"d2a9bf9f9b8cd012e4d8789f8c5b5f0f0eb339ecdb3c505e82c537a4727ec528"} Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.042159 4848 scope.go:117] "RemoveContainer" containerID="14f013eceadc2ad0e7111f2fe485eefe5cb5a76b319302c45f850d5966f830ea" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.042317 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.060381 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-web-config\") pod \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.060450 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-thanos-prometheus-http-client-file\") pod \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.060477 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.060589 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-prometheus-metric-storage-rulefiles-0\") pod \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.060634 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cmxf\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-kube-api-access-8cmxf\") pod \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.060651 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config\") pod \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.060711 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config-out\") pod \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.060745 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-tls-assets\") pod \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\" (UID: \"b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1\") " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.062865 4848 generic.go:334] "Generic (PLEG): container finished" podID="565e9c18-f099-49a4-9457-725b85e0cb7a" containerID="def06344ecd9560b993c752d7c61cedd7d29be1747532632fd15618959d6afaf" exitCode=0 Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.062912 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" event={"ID":"565e9c18-f099-49a4-9457-725b85e0cb7a","Type":"ContainerDied","Data":"def06344ecd9560b993c752d7c61cedd7d29be1747532632fd15618959d6afaf"} Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.065398 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" (UID: "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.075922 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" (UID: "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.090654 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-kube-api-access-8cmxf" (OuterVolumeSpecName: "kube-api-access-8cmxf") pod "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" (UID: "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1"). InnerVolumeSpecName "kube-api-access-8cmxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.092143 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" (UID: "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.096386 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config" (OuterVolumeSpecName: "config") pod "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" (UID: "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.102485 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config-out" (OuterVolumeSpecName: "config-out") pod "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" (UID: "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.102779 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" (UID: "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.133881 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-web-config" (OuterVolumeSpecName: "web-config") pod "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" (UID: "b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.149510 4848 scope.go:117] "RemoveContainer" containerID="4a29902bc73018a50387e59b45b3c5a4eec4d42ef6f1c00e1528e361cd96a8c4" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.162643 4848 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.163085 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cmxf\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-kube-api-access-8cmxf\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.163155 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.163243 4848 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-config-out\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.163297 4848 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.163345 4848 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-web-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.163394 4848 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.163463 4848 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.204166 4848 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.253170 4848 scope.go:117] "RemoveContainer" containerID="57084ce6e036e4bb452a513c26295d75c4c3ba17caf688ccffdc8a807ef9505e" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.265090 4848 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.460851 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.461204 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.481032 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:11:20 crc kubenswrapper[4848]: E1204 14:11:20.481663 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="config-reloader" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.481687 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="config-reloader" Dec 04 14:11:20 crc kubenswrapper[4848]: E1204 14:11:20.481726 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="thanos-sidecar" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.481735 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="thanos-sidecar" Dec 04 14:11:20 crc kubenswrapper[4848]: E1204 14:11:20.481753 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="prometheus" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.481762 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="prometheus" Dec 04 14:11:20 crc kubenswrapper[4848]: E1204 14:11:20.481778 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="init-config-reloader" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.481787 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="init-config-reloader" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.482072 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="prometheus" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.482096 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="config-reloader" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.482107 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="thanos-sidecar" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.488104 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.494855 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.495400 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.495696 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.496002 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.496831 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-dzcpz" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.502865 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.505044 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.505927 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.576477 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.576529 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.576570 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhk5s\" (UniqueName: \"kubernetes.io/projected/a9e4a10a-a31e-431d-99e2-d4205440fad6-kube-api-access-dhk5s\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.576656 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a9e4a10a-a31e-431d-99e2-d4205440fad6-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.576845 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-config\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.576883 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.576933 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.577350 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.577393 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a9e4a10a-a31e-431d-99e2-d4205440fad6-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.577421 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.577488 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a9e4a10a-a31e-431d-99e2-d4205440fad6-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.598022 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-61ea-account-create-update-mtrsm"] Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.679166 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-config\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.679248 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.679453 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680088 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680122 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a9e4a10a-a31e-431d-99e2-d4205440fad6-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680142 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680187 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a9e4a10a-a31e-431d-99e2-d4205440fad6-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680240 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680259 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680285 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhk5s\" (UniqueName: \"kubernetes.io/projected/a9e4a10a-a31e-431d-99e2-d4205440fad6-kube-api-access-dhk5s\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680328 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a9e4a10a-a31e-431d-99e2-d4205440fad6-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.680341 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.682390 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a9e4a10a-a31e-431d-99e2-d4205440fad6-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.686928 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-config\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.687819 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.691683 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.692119 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.693802 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.694860 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a9e4a10a-a31e-431d-99e2-d4205440fad6-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.695796 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a9e4a10a-a31e-431d-99e2-d4205440fad6-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.700717 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a9e4a10a-a31e-431d-99e2-d4205440fad6-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.701633 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhk5s\" (UniqueName: \"kubernetes.io/projected/a9e4a10a-a31e-431d-99e2-d4205440fad6-kube-api-access-dhk5s\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.837035 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"a9e4a10a-a31e-431d-99e2-d4205440fad6\") " pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:20 crc kubenswrapper[4848]: I1204 14:11:20.851667 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.162201 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.162460 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cxgrr"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.217128 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-cpcmr" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.263046 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-ldg2g"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.280417 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-8pzdg"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.434925 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rg7xl-config-74z5m"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.436655 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.440714 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.473352 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rg7xl-config-74z5m"] Dec 04 14:11:21 crc kubenswrapper[4848]: W1204 14:11:21.508000 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e722893_4a07_44e8_9e5a_3e04ff50dc47.slice/crio-250ad5dd1b248fff8d58f8caf941db74306c0b13accccf8efedad9bc6199bdc7 WatchSource:0}: Error finding container 250ad5dd1b248fff8d58f8caf941db74306c0b13accccf8efedad9bc6199bdc7: Status 404 returned error can't find the container with id 250ad5dd1b248fff8d58f8caf941db74306c0b13accccf8efedad9bc6199bdc7 Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.515584 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-additional-scripts\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.515634 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-scripts\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.515667 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-log-ovn\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.515737 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run-ovn\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.515805 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kftxf\" (UniqueName: \"kubernetes.io/projected/87c6c058-c395-40c9-b6f1-b1a0580447b5-kube-api-access-kftxf\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.516042 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.545709 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2804-account-create-update-mrhhd"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.574109 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-021e-account-create-update-g264j"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.598247 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-fd58-account-create-update-r8csb"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.612512 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.618659 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.618790 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-additional-scripts\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.618824 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-scripts\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.618849 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-log-ovn\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.618913 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run-ovn\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.618978 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kftxf\" (UniqueName: \"kubernetes.io/projected/87c6c058-c395-40c9-b6f1-b1a0580447b5-kube-api-access-kftxf\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.619648 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.620644 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-log-ovn\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.623758 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-scripts\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.624313 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run-ovn\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.624418 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-dchv8"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.625458 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-additional-scripts\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.629885 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s2z5f"] Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.644644 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.673157 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kftxf\" (UniqueName: \"kubernetes.io/projected/87c6c058-c395-40c9-b6f1-b1a0580447b5-kube-api-access-kftxf\") pod \"ovn-controller-rg7xl-config-74z5m\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.720227 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk7w4\" (UniqueName: \"kubernetes.io/projected/565e9c18-f099-49a4-9457-725b85e0cb7a-kube-api-access-vk7w4\") pod \"565e9c18-f099-49a4-9457-725b85e0cb7a\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.720536 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-dns-svc\") pod \"565e9c18-f099-49a4-9457-725b85e0cb7a\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.720576 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-config\") pod \"565e9c18-f099-49a4-9457-725b85e0cb7a\" (UID: \"565e9c18-f099-49a4-9457-725b85e0cb7a\") " Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.735824 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565e9c18-f099-49a4-9457-725b85e0cb7a-kube-api-access-vk7w4" (OuterVolumeSpecName: "kube-api-access-vk7w4") pod "565e9c18-f099-49a4-9457-725b85e0cb7a" (UID: "565e9c18-f099-49a4-9457-725b85e0cb7a"). InnerVolumeSpecName "kube-api-access-vk7w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.770502 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.815909 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "565e9c18-f099-49a4-9457-725b85e0cb7a" (UID: "565e9c18-f099-49a4-9457-725b85e0cb7a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.824483 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.824521 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk7w4\" (UniqueName: \"kubernetes.io/projected/565e9c18-f099-49a4-9457-725b85e0cb7a-kube-api-access-vk7w4\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.860591 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-config" (OuterVolumeSpecName: "config") pod "565e9c18-f099-49a4-9457-725b85e0cb7a" (UID: "565e9c18-f099-49a4-9457-725b85e0cb7a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:21 crc kubenswrapper[4848]: I1204 14:11:21.926340 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/565e9c18-f099-49a4-9457-725b85e0cb7a-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.124850 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ldg2g" event={"ID":"4e722893-4a07-44e8-9e5a-3e04ff50dc47","Type":"ContainerStarted","Data":"250ad5dd1b248fff8d58f8caf941db74306c0b13accccf8efedad9bc6199bdc7"} Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.126280 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cxgrr" event={"ID":"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f","Type":"ContainerStarted","Data":"5875389c91ad122abf88c206eb7623756b9577a9ce01c87023ed3bb10e36cf7a"} Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.131737 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.137261 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-2hkgw" event={"ID":"565e9c18-f099-49a4-9457-725b85e0cb7a","Type":"ContainerDied","Data":"f5d8714e32ed12c5b1638fef3415452f8a9e50add47bdd61a4ebe0af7b09c532"} Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.173562 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-2hkgw"] Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.183076 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-2hkgw"] Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.411233 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="565e9c18-f099-49a4-9457-725b85e0cb7a" path="/var/lib/kubelet/pods/565e9c18-f099-49a4-9457-725b85e0cb7a/volumes" Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.412238 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" path="/var/lib/kubelet/pods/b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1/volumes" Dec 04 14:11:22 crc kubenswrapper[4848]: I1204 14:11:22.477340 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="b8f17c88-ddc6-4ee8-8834-cf1157c4cfa1" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.136:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:11:24 crc kubenswrapper[4848]: I1204 14:11:24.993162 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 04 14:11:29 crc kubenswrapper[4848]: W1204 14:11:29.642045 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf74ef9c_2afd_4236_901e_82c48594f70f.slice/crio-7056def443aefa5a93c1ab43e3516d604a92feb2f580388184cd7d23d89916f3 WatchSource:0}: Error finding container 7056def443aefa5a93c1ab43e3516d604a92feb2f580388184cd7d23d89916f3: Status 404 returned error can't find the container with id 7056def443aefa5a93c1ab43e3516d604a92feb2f580388184cd7d23d89916f3 Dec 04 14:11:29 crc kubenswrapper[4848]: W1204 14:11:29.645024 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf2c3a14_5cfd_49a1_9337_fb0b61611669.slice/crio-cbedabd2535519f9c4c659914407f702e88db180f49b973ff04fbcc9b7f98053 WatchSource:0}: Error finding container cbedabd2535519f9c4c659914407f702e88db180f49b973ff04fbcc9b7f98053: Status 404 returned error can't find the container with id cbedabd2535519f9c4c659914407f702e88db180f49b973ff04fbcc9b7f98053 Dec 04 14:11:29 crc kubenswrapper[4848]: I1204 14:11:29.653829 4848 scope.go:117] "RemoveContainer" containerID="82a8f10da12cd1dca25f294fd60d46653a7a4b4a0c6c4c7c22457cd12570dea7" Dec 04 14:11:29 crc kubenswrapper[4848]: W1204 14:11:29.657898 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a756447_5ce8_4b33_9075_e3b6100d807c.slice/crio-2f839d23691da9612b6a6a7244aeaaee3c5e87f328f4c28281a5574e4471af74 WatchSource:0}: Error finding container 2f839d23691da9612b6a6a7244aeaaee3c5e87f328f4c28281a5574e4471af74: Status 404 returned error can't find the container with id 2f839d23691da9612b6a6a7244aeaaee3c5e87f328f4c28281a5574e4471af74 Dec 04 14:11:29 crc kubenswrapper[4848]: W1204 14:11:29.699687 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcac2310b_905a_42df_92d8_611fa944c8f6.slice/crio-30e167c3c7b934c555017965ccb2a4065ed0b7db8341a76ef6238a8a7d0e6f0c WatchSource:0}: Error finding container 30e167c3c7b934c555017965ccb2a4065ed0b7db8341a76ef6238a8a7d0e6f0c: Status 404 returned error can't find the container with id 30e167c3c7b934c555017965ccb2a4065ed0b7db8341a76ef6238a8a7d0e6f0c Dec 04 14:11:29 crc kubenswrapper[4848]: W1204 14:11:29.740071 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1d785c9_6be8_49a5_a4f6_b20f8c93b1ec.slice/crio-ae5ad07cfde5bf13f698e48521145e2e489e10777d128b8cbe87ccf3de03c79f WatchSource:0}: Error finding container ae5ad07cfde5bf13f698e48521145e2e489e10777d128b8cbe87ccf3de03c79f: Status 404 returned error can't find the container with id ae5ad07cfde5bf13f698e48521145e2e489e10777d128b8cbe87ccf3de03c79f Dec 04 14:11:29 crc kubenswrapper[4848]: I1204 14:11:29.845786 4848 scope.go:117] "RemoveContainer" containerID="def06344ecd9560b993c752d7c61cedd7d29be1747532632fd15618959d6afaf" Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.040507 4848 scope.go:117] "RemoveContainer" containerID="66e54b150809137932b9d4b609fafb607f0bc4d29ae529f038da48b097ecbd3c" Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.229819 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2804-account-create-update-mrhhd" event={"ID":"6a756447-5ce8-4b33-9075-e3b6100d807c","Type":"ContainerStarted","Data":"391a991f094c6871eb4b127ade42b2169669f19a1271921c0e6fd15574eca231"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.229885 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2804-account-create-update-mrhhd" event={"ID":"6a756447-5ce8-4b33-9075-e3b6100d807c","Type":"ContainerStarted","Data":"2f839d23691da9612b6a6a7244aeaaee3c5e87f328f4c28281a5574e4471af74"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.231325 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s2z5f" event={"ID":"cac2310b-905a-42df-92d8-611fa944c8f6","Type":"ContainerStarted","Data":"30e167c3c7b934c555017965ccb2a4065ed0b7db8341a76ef6238a8a7d0e6f0c"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.236086 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-61ea-account-create-update-mtrsm" event={"ID":"df74ef9c-2afd-4236-901e-82c48594f70f","Type":"ContainerStarted","Data":"7056def443aefa5a93c1ab43e3516d604a92feb2f580388184cd7d23d89916f3"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.237438 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-8pzdg" event={"ID":"cf2c3a14-5cfd-49a1-9337-fb0b61611669","Type":"ContainerStarted","Data":"cbedabd2535519f9c4c659914407f702e88db180f49b973ff04fbcc9b7f98053"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.251433 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fd58-account-create-update-r8csb" event={"ID":"41c0df85-a096-4e29-9dae-4e98007e43ff","Type":"ContainerStarted","Data":"cd29192b1c2a802271aabaa102cef3111d8437b1076fd38e2fa8a071afd97602"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.253815 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-021e-account-create-update-g264j" event={"ID":"867de14d-11f1-4306-873d-cdc3f6f90216","Type":"ContainerStarted","Data":"7ac88be9572da37b1b8a8c385b55df67ff7c71dd2652ad984c006cc6a6a5f5a6"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.255636 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-dchv8" event={"ID":"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec","Type":"ContainerStarted","Data":"ae5ad07cfde5bf13f698e48521145e2e489e10777d128b8cbe87ccf3de03c79f"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.259826 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-2804-account-create-update-mrhhd" podStartSLOduration=12.259803198 podStartE2EDuration="12.259803198s" podCreationTimestamp="2025-12-04 14:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:30.249761655 +0000 UTC m=+1394.192258193" watchObservedRunningTime="2025-12-04 14:11:30.259803198 +0000 UTC m=+1394.202299726" Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.260406 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"d2af561d-8819-4175-89c0-4a65a661fdb6","Type":"ContainerStarted","Data":"c97d3f879a997959064e5310eb04bb6c1cbf55a4328493f9ab33cec3d34900f8"} Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.280196 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.385773 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rg7xl-config-74z5m"] Dec 04 14:11:30 crc kubenswrapper[4848]: W1204 14:11:30.396643 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87c6c058_c395_40c9_b6f1_b1a0580447b5.slice/crio-dfa8215fe221c1b97c4839589c1c18471d33762822e4f63d8f287a0f4574664b WatchSource:0}: Error finding container dfa8215fe221c1b97c4839589c1c18471d33762822e4f63d8f287a0f4574664b: Status 404 returned error can't find the container with id dfa8215fe221c1b97c4839589c1c18471d33762822e4f63d8f287a0f4574664b Dec 04 14:11:30 crc kubenswrapper[4848]: I1204 14:11:30.992279 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-rg7xl" Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.273670 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ldg2g" event={"ID":"4e722893-4a07-44e8-9e5a-3e04ff50dc47","Type":"ContainerStarted","Data":"a68eff2e9b3822442aff1993cc6d0135e5d3bbac4dc7536797a5c55789a98e01"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.277468 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rg7xl-config-74z5m" event={"ID":"87c6c058-c395-40c9-b6f1-b1a0580447b5","Type":"ContainerStarted","Data":"e8e23d771c6fa0161054cefb3e13fc1e598648961438525837ca0dced9b5ddcf"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.277517 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rg7xl-config-74z5m" event={"ID":"87c6c058-c395-40c9-b6f1-b1a0580447b5","Type":"ContainerStarted","Data":"dfa8215fe221c1b97c4839589c1c18471d33762822e4f63d8f287a0f4574664b"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.279307 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s2z5f" event={"ID":"cac2310b-905a-42df-92d8-611fa944c8f6","Type":"ContainerStarted","Data":"1d2280872e954728b571252f7e9904fc32854a28826a1e941433fbeb18ae94ba"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.281734 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-61ea-account-create-update-mtrsm" event={"ID":"df74ef9c-2afd-4236-901e-82c48594f70f","Type":"ContainerStarted","Data":"2728ee54f8ea9c9f2e08c4ed603b6ddb8e56b9211c9c2f1a3db03090976d8cc9"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.291311 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-dchv8" event={"ID":"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec","Type":"ContainerStarted","Data":"5e9d418ccd6305d98dee61226d0bb4de9c1d0061f7bec7ab92a5ae7840ba2398"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.297911 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cxgrr" event={"ID":"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f","Type":"ContainerStarted","Data":"5bf60c5607da1252b532d01b4a4deee702442142969cf45cc40dc08341cee54e"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.298664 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-ldg2g" podStartSLOduration=13.298649844 podStartE2EDuration="13.298649844s" podCreationTimestamp="2025-12-04 14:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:31.290578869 +0000 UTC m=+1395.233075397" watchObservedRunningTime="2025-12-04 14:11:31.298649844 +0000 UTC m=+1395.241146372" Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.308579 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"b63b762daf19a333393bc42a750f10199f625cc4e97fe97e45dd9fa7f4441a52"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.308631 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"0224ee2a48223c88f8361a124c047a64796fe8afa03856c41612ed3da4dad976"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.308902 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-s2z5f" podStartSLOduration=12.308888893 podStartE2EDuration="12.308888893s" podCreationTimestamp="2025-12-04 14:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:31.307851768 +0000 UTC m=+1395.250348296" watchObservedRunningTime="2025-12-04 14:11:31.308888893 +0000 UTC m=+1395.251385431" Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.310089 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a9e4a10a-a31e-431d-99e2-d4205440fad6","Type":"ContainerStarted","Data":"e03b6616aca32496e1e270f285552cbb8d95464814cfae2c607be6f491e64281"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.311963 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r7w8t" event={"ID":"9925991d-d748-4a39-96a0-3bf2e391cad7","Type":"ContainerStarted","Data":"2ded7473b5d4b3564f6fc47d069bdb2ecb6bd7939ef27e3dfe27f9848e205a6a"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.317241 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fd58-account-create-update-r8csb" event={"ID":"41c0df85-a096-4e29-9dae-4e98007e43ff","Type":"ContainerStarted","Data":"5b709be838cafa6f4d427b1f3e0a3be519e9629525ded45e7ff845d623dd9a82"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.320642 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-021e-account-create-update-g264j" event={"ID":"867de14d-11f1-4306-873d-cdc3f6f90216","Type":"ContainerStarted","Data":"93f0d85a257f7eb16a75dec0e34adf2c4206b616299aaa2ff15080fde79aa7fc"} Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.353970 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-61ea-account-create-update-mtrsm" podStartSLOduration=13.353937416 podStartE2EDuration="13.353937416s" podCreationTimestamp="2025-12-04 14:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:31.34090366 +0000 UTC m=+1395.283400188" watchObservedRunningTime="2025-12-04 14:11:31.353937416 +0000 UTC m=+1395.296433944" Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.378210 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-r7w8t" podStartSLOduration=2.822610699 podStartE2EDuration="20.378192804s" podCreationTimestamp="2025-12-04 14:11:11 +0000 UTC" firstStartedPulling="2025-12-04 14:11:12.414392433 +0000 UTC m=+1376.356888961" lastFinishedPulling="2025-12-04 14:11:29.969974538 +0000 UTC m=+1393.912471066" observedRunningTime="2025-12-04 14:11:31.358006305 +0000 UTC m=+1395.300502833" watchObservedRunningTime="2025-12-04 14:11:31.378192804 +0000 UTC m=+1395.320689332" Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.398468 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-fd58-account-create-update-r8csb" podStartSLOduration=13.397354019 podStartE2EDuration="13.397354019s" podCreationTimestamp="2025-12-04 14:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:31.37059061 +0000 UTC m=+1395.313087138" watchObservedRunningTime="2025-12-04 14:11:31.397354019 +0000 UTC m=+1395.339850557" Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.408412 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-021e-account-create-update-g264j" podStartSLOduration=12.408397327 podStartE2EDuration="12.408397327s" podCreationTimestamp="2025-12-04 14:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:31.381512075 +0000 UTC m=+1395.324008613" watchObservedRunningTime="2025-12-04 14:11:31.408397327 +0000 UTC m=+1395.350893855" Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.420108 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-cxgrr" podStartSLOduration=13.42009065 podStartE2EDuration="13.42009065s" podCreationTimestamp="2025-12-04 14:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:31.394672514 +0000 UTC m=+1395.337169052" watchObservedRunningTime="2025-12-04 14:11:31.42009065 +0000 UTC m=+1395.362587178" Dec 04 14:11:31 crc kubenswrapper[4848]: I1204 14:11:31.423444 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-dchv8" podStartSLOduration=13.423436392 podStartE2EDuration="13.423436392s" podCreationTimestamp="2025-12-04 14:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:31.412163838 +0000 UTC m=+1395.354660386" watchObservedRunningTime="2025-12-04 14:11:31.423436392 +0000 UTC m=+1395.365932920" Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.337293 4848 generic.go:334] "Generic (PLEG): container finished" podID="41c0df85-a096-4e29-9dae-4e98007e43ff" containerID="5b709be838cafa6f4d427b1f3e0a3be519e9629525ded45e7ff845d623dd9a82" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.337371 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fd58-account-create-update-r8csb" event={"ID":"41c0df85-a096-4e29-9dae-4e98007e43ff","Type":"ContainerDied","Data":"5b709be838cafa6f4d427b1f3e0a3be519e9629525ded45e7ff845d623dd9a82"} Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.341029 4848 generic.go:334] "Generic (PLEG): container finished" podID="4e722893-4a07-44e8-9e5a-3e04ff50dc47" containerID="a68eff2e9b3822442aff1993cc6d0135e5d3bbac4dc7536797a5c55789a98e01" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.341147 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ldg2g" event={"ID":"4e722893-4a07-44e8-9e5a-3e04ff50dc47","Type":"ContainerDied","Data":"a68eff2e9b3822442aff1993cc6d0135e5d3bbac4dc7536797a5c55789a98e01"} Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.343526 4848 generic.go:334] "Generic (PLEG): container finished" podID="87c6c058-c395-40c9-b6f1-b1a0580447b5" containerID="e8e23d771c6fa0161054cefb3e13fc1e598648961438525837ca0dced9b5ddcf" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.343587 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rg7xl-config-74z5m" event={"ID":"87c6c058-c395-40c9-b6f1-b1a0580447b5","Type":"ContainerDied","Data":"e8e23d771c6fa0161054cefb3e13fc1e598648961438525837ca0dced9b5ddcf"} Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.345847 4848 generic.go:334] "Generic (PLEG): container finished" podID="867de14d-11f1-4306-873d-cdc3f6f90216" containerID="93f0d85a257f7eb16a75dec0e34adf2c4206b616299aaa2ff15080fde79aa7fc" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.345890 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-021e-account-create-update-g264j" event={"ID":"867de14d-11f1-4306-873d-cdc3f6f90216","Type":"ContainerDied","Data":"93f0d85a257f7eb16a75dec0e34adf2c4206b616299aaa2ff15080fde79aa7fc"} Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.352346 4848 generic.go:334] "Generic (PLEG): container finished" podID="df74ef9c-2afd-4236-901e-82c48594f70f" containerID="2728ee54f8ea9c9f2e08c4ed603b6ddb8e56b9211c9c2f1a3db03090976d8cc9" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.352410 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-61ea-account-create-update-mtrsm" event={"ID":"df74ef9c-2afd-4236-901e-82c48594f70f","Type":"ContainerDied","Data":"2728ee54f8ea9c9f2e08c4ed603b6ddb8e56b9211c9c2f1a3db03090976d8cc9"} Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.369372 4848 generic.go:334] "Generic (PLEG): container finished" podID="a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec" containerID="5e9d418ccd6305d98dee61226d0bb4de9c1d0061f7bec7ab92a5ae7840ba2398" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.369437 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-dchv8" event={"ID":"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec","Type":"ContainerDied","Data":"5e9d418ccd6305d98dee61226d0bb4de9c1d0061f7bec7ab92a5ae7840ba2398"} Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.371734 4848 generic.go:334] "Generic (PLEG): container finished" podID="97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f" containerID="5bf60c5607da1252b532d01b4a4deee702442142969cf45cc40dc08341cee54e" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.371801 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cxgrr" event={"ID":"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f","Type":"ContainerDied","Data":"5bf60c5607da1252b532d01b4a4deee702442142969cf45cc40dc08341cee54e"} Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.374414 4848 generic.go:334] "Generic (PLEG): container finished" podID="6a756447-5ce8-4b33-9075-e3b6100d807c" containerID="391a991f094c6871eb4b127ade42b2169669f19a1271921c0e6fd15574eca231" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.374484 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2804-account-create-update-mrhhd" event={"ID":"6a756447-5ce8-4b33-9075-e3b6100d807c","Type":"ContainerDied","Data":"391a991f094c6871eb4b127ade42b2169669f19a1271921c0e6fd15574eca231"} Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.376160 4848 generic.go:334] "Generic (PLEG): container finished" podID="cac2310b-905a-42df-92d8-611fa944c8f6" containerID="1d2280872e954728b571252f7e9904fc32854a28826a1e941433fbeb18ae94ba" exitCode=0 Dec 04 14:11:32 crc kubenswrapper[4848]: I1204 14:11:32.377397 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s2z5f" event={"ID":"cac2310b-905a-42df-92d8-611fa944c8f6","Type":"ContainerDied","Data":"1d2280872e954728b571252f7e9904fc32854a28826a1e941433fbeb18ae94ba"} Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.024644 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.125578 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run\") pod \"87c6c058-c395-40c9-b6f1-b1a0580447b5\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.125672 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run" (OuterVolumeSpecName: "var-run") pod "87c6c058-c395-40c9-b6f1-b1a0580447b5" (UID: "87c6c058-c395-40c9-b6f1-b1a0580447b5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.125979 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-scripts\") pod \"87c6c058-c395-40c9-b6f1-b1a0580447b5\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126014 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-additional-scripts\") pod \"87c6c058-c395-40c9-b6f1-b1a0580447b5\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126064 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-log-ovn\") pod \"87c6c058-c395-40c9-b6f1-b1a0580447b5\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126133 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run-ovn\") pod \"87c6c058-c395-40c9-b6f1-b1a0580447b5\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126175 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kftxf\" (UniqueName: \"kubernetes.io/projected/87c6c058-c395-40c9-b6f1-b1a0580447b5-kube-api-access-kftxf\") pod \"87c6c058-c395-40c9-b6f1-b1a0580447b5\" (UID: \"87c6c058-c395-40c9-b6f1-b1a0580447b5\") " Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126314 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "87c6c058-c395-40c9-b6f1-b1a0580447b5" (UID: "87c6c058-c395-40c9-b6f1-b1a0580447b5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126300 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "87c6c058-c395-40c9-b6f1-b1a0580447b5" (UID: "87c6c058-c395-40c9-b6f1-b1a0580447b5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126697 4848 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126717 4848 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.126741 4848 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87c6c058-c395-40c9-b6f1-b1a0580447b5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.127017 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "87c6c058-c395-40c9-b6f1-b1a0580447b5" (UID: "87c6c058-c395-40c9-b6f1-b1a0580447b5"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.127107 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-scripts" (OuterVolumeSpecName: "scripts") pod "87c6c058-c395-40c9-b6f1-b1a0580447b5" (UID: "87c6c058-c395-40c9-b6f1-b1a0580447b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.130991 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87c6c058-c395-40c9-b6f1-b1a0580447b5-kube-api-access-kftxf" (OuterVolumeSpecName: "kube-api-access-kftxf") pod "87c6c058-c395-40c9-b6f1-b1a0580447b5" (UID: "87c6c058-c395-40c9-b6f1-b1a0580447b5"). InnerVolumeSpecName "kube-api-access-kftxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.228342 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kftxf\" (UniqueName: \"kubernetes.io/projected/87c6c058-c395-40c9-b6f1-b1a0580447b5-kube-api-access-kftxf\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.228384 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.228396 4848 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/87c6c058-c395-40c9-b6f1-b1a0580447b5-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.412634 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a9e4a10a-a31e-431d-99e2-d4205440fad6","Type":"ContainerStarted","Data":"525d3f55d89a39b18bfc722d1462ff7c6d9b08a8118d7c1217aeabc268276191"} Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.414830 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rg7xl-config-74z5m" event={"ID":"87c6c058-c395-40c9-b6f1-b1a0580447b5","Type":"ContainerDied","Data":"dfa8215fe221c1b97c4839589c1c18471d33762822e4f63d8f287a0f4574664b"} Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.414860 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfa8215fe221c1b97c4839589c1c18471d33762822e4f63d8f287a0f4574664b" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.414907 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rg7xl-config-74z5m" Dec 04 14:11:33 crc kubenswrapper[4848]: I1204 14:11:33.420833 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"110a9c978ddc6712bdfcc9d4376bccac383fee19b436fa229c46cf7ab4afbbbf"} Dec 04 14:11:34 crc kubenswrapper[4848]: I1204 14:11:34.200235 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rg7xl-config-74z5m"] Dec 04 14:11:34 crc kubenswrapper[4848]: I1204 14:11:34.212222 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rg7xl-config-74z5m"] Dec 04 14:11:34 crc kubenswrapper[4848]: I1204 14:11:34.408380 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87c6c058-c395-40c9-b6f1-b1a0580447b5" path="/var/lib/kubelet/pods/87c6c058-c395-40c9-b6f1-b1a0580447b5/volumes" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.455920 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cxgrr" event={"ID":"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f","Type":"ContainerDied","Data":"5875389c91ad122abf88c206eb7623756b9577a9ce01c87023ed3bb10e36cf7a"} Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.456688 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5875389c91ad122abf88c206eb7623756b9577a9ce01c87023ed3bb10e36cf7a" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.458784 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fd58-account-create-update-r8csb" event={"ID":"41c0df85-a096-4e29-9dae-4e98007e43ff","Type":"ContainerDied","Data":"cd29192b1c2a802271aabaa102cef3111d8437b1076fd38e2fa8a071afd97602"} Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.458813 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd29192b1c2a802271aabaa102cef3111d8437b1076fd38e2fa8a071afd97602" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.461117 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ldg2g" event={"ID":"4e722893-4a07-44e8-9e5a-3e04ff50dc47","Type":"ContainerDied","Data":"250ad5dd1b248fff8d58f8caf941db74306c0b13accccf8efedad9bc6199bdc7"} Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.461166 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="250ad5dd1b248fff8d58f8caf941db74306c0b13accccf8efedad9bc6199bdc7" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.462791 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-021e-account-create-update-g264j" event={"ID":"867de14d-11f1-4306-873d-cdc3f6f90216","Type":"ContainerDied","Data":"7ac88be9572da37b1b8a8c385b55df67ff7c71dd2652ad984c006cc6a6a5f5a6"} Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.462810 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ac88be9572da37b1b8a8c385b55df67ff7c71dd2652ad984c006cc6a6a5f5a6" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.464527 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2804-account-create-update-mrhhd" event={"ID":"6a756447-5ce8-4b33-9075-e3b6100d807c","Type":"ContainerDied","Data":"2f839d23691da9612b6a6a7244aeaaee3c5e87f328f4c28281a5574e4471af74"} Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.464563 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f839d23691da9612b6a6a7244aeaaee3c5e87f328f4c28281a5574e4471af74" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.466215 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s2z5f" event={"ID":"cac2310b-905a-42df-92d8-611fa944c8f6","Type":"ContainerDied","Data":"30e167c3c7b934c555017965ccb2a4065ed0b7db8341a76ef6238a8a7d0e6f0c"} Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.466257 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30e167c3c7b934c555017965ccb2a4065ed0b7db8341a76ef6238a8a7d0e6f0c" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.468071 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-61ea-account-create-update-mtrsm" event={"ID":"df74ef9c-2afd-4236-901e-82c48594f70f","Type":"ContainerDied","Data":"7056def443aefa5a93c1ab43e3516d604a92feb2f580388184cd7d23d89916f3"} Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.468098 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7056def443aefa5a93c1ab43e3516d604a92feb2f580388184cd7d23d89916f3" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.470783 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-dchv8" event={"ID":"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec","Type":"ContainerDied","Data":"ae5ad07cfde5bf13f698e48521145e2e489e10777d128b8cbe87ccf3de03c79f"} Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.470834 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae5ad07cfde5bf13f698e48521145e2e489e10777d128b8cbe87ccf3de03c79f" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.578537 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.627802 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.628290 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.649027 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-dchv8" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.690617 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.693175 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.701473 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-operator-scripts\") pod \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\" (UID: \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.701511 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c742l\" (UniqueName: \"kubernetes.io/projected/41c0df85-a096-4e29-9dae-4e98007e43ff-kube-api-access-c742l\") pod \"41c0df85-a096-4e29-9dae-4e98007e43ff\" (UID: \"41c0df85-a096-4e29-9dae-4e98007e43ff\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.701655 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsw2m\" (UniqueName: \"kubernetes.io/projected/cac2310b-905a-42df-92d8-611fa944c8f6-kube-api-access-tsw2m\") pod \"cac2310b-905a-42df-92d8-611fa944c8f6\" (UID: \"cac2310b-905a-42df-92d8-611fa944c8f6\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.701747 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41c0df85-a096-4e29-9dae-4e98007e43ff-operator-scripts\") pod \"41c0df85-a096-4e29-9dae-4e98007e43ff\" (UID: \"41c0df85-a096-4e29-9dae-4e98007e43ff\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.701824 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjjgb\" (UniqueName: \"kubernetes.io/projected/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-kube-api-access-tjjgb\") pod \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\" (UID: \"a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.701868 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/867de14d-11f1-4306-873d-cdc3f6f90216-operator-scripts\") pod \"867de14d-11f1-4306-873d-cdc3f6f90216\" (UID: \"867de14d-11f1-4306-873d-cdc3f6f90216\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.701939 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c52ht\" (UniqueName: \"kubernetes.io/projected/867de14d-11f1-4306-873d-cdc3f6f90216-kube-api-access-c52ht\") pod \"867de14d-11f1-4306-873d-cdc3f6f90216\" (UID: \"867de14d-11f1-4306-873d-cdc3f6f90216\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.702004 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cac2310b-905a-42df-92d8-611fa944c8f6-operator-scripts\") pod \"cac2310b-905a-42df-92d8-611fa944c8f6\" (UID: \"cac2310b-905a-42df-92d8-611fa944c8f6\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.702834 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cac2310b-905a-42df-92d8-611fa944c8f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cac2310b-905a-42df-92d8-611fa944c8f6" (UID: "cac2310b-905a-42df-92d8-611fa944c8f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.703220 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/867de14d-11f1-4306-873d-cdc3f6f90216-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "867de14d-11f1-4306-873d-cdc3f6f90216" (UID: "867de14d-11f1-4306-873d-cdc3f6f90216"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.703779 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec" (UID: "a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.704254 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41c0df85-a096-4e29-9dae-4e98007e43ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "41c0df85-a096-4e29-9dae-4e98007e43ff" (UID: "41c0df85-a096-4e29-9dae-4e98007e43ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.705599 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.719872 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41c0df85-a096-4e29-9dae-4e98007e43ff-kube-api-access-c742l" (OuterVolumeSpecName: "kube-api-access-c742l") pod "41c0df85-a096-4e29-9dae-4e98007e43ff" (UID: "41c0df85-a096-4e29-9dae-4e98007e43ff"). InnerVolumeSpecName "kube-api-access-c742l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.722039 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-kube-api-access-tjjgb" (OuterVolumeSpecName: "kube-api-access-tjjgb") pod "a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec" (UID: "a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec"). InnerVolumeSpecName "kube-api-access-tjjgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.733725 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.740454 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/867de14d-11f1-4306-873d-cdc3f6f90216-kube-api-access-c52ht" (OuterVolumeSpecName: "kube-api-access-c52ht") pod "867de14d-11f1-4306-873d-cdc3f6f90216" (UID: "867de14d-11f1-4306-873d-cdc3f6f90216"). InnerVolumeSpecName "kube-api-access-c52ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.749861 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cac2310b-905a-42df-92d8-611fa944c8f6-kube-api-access-tsw2m" (OuterVolumeSpecName: "kube-api-access-tsw2m") pod "cac2310b-905a-42df-92d8-611fa944c8f6" (UID: "cac2310b-905a-42df-92d8-611fa944c8f6"). InnerVolumeSpecName "kube-api-access-tsw2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803183 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l44fp\" (UniqueName: \"kubernetes.io/projected/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-kube-api-access-l44fp\") pod \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\" (UID: \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803232 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a756447-5ce8-4b33-9075-e3b6100d807c-operator-scripts\") pod \"6a756447-5ce8-4b33-9075-e3b6100d807c\" (UID: \"6a756447-5ce8-4b33-9075-e3b6100d807c\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803278 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfhjr\" (UniqueName: \"kubernetes.io/projected/df74ef9c-2afd-4236-901e-82c48594f70f-kube-api-access-wfhjr\") pod \"df74ef9c-2afd-4236-901e-82c48594f70f\" (UID: \"df74ef9c-2afd-4236-901e-82c48594f70f\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803352 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e722893-4a07-44e8-9e5a-3e04ff50dc47-operator-scripts\") pod \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\" (UID: \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803436 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh27s\" (UniqueName: \"kubernetes.io/projected/4e722893-4a07-44e8-9e5a-3e04ff50dc47-kube-api-access-zh27s\") pod \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\" (UID: \"4e722893-4a07-44e8-9e5a-3e04ff50dc47\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803463 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df74ef9c-2afd-4236-901e-82c48594f70f-operator-scripts\") pod \"df74ef9c-2afd-4236-901e-82c48594f70f\" (UID: \"df74ef9c-2afd-4236-901e-82c48594f70f\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803480 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-operator-scripts\") pod \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\" (UID: \"97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803509 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njm48\" (UniqueName: \"kubernetes.io/projected/6a756447-5ce8-4b33-9075-e3b6100d807c-kube-api-access-njm48\") pod \"6a756447-5ce8-4b33-9075-e3b6100d807c\" (UID: \"6a756447-5ce8-4b33-9075-e3b6100d807c\") " Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803969 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c52ht\" (UniqueName: \"kubernetes.io/projected/867de14d-11f1-4306-873d-cdc3f6f90216-kube-api-access-c52ht\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803986 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cac2310b-905a-42df-92d8-611fa944c8f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.803996 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.804006 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c742l\" (UniqueName: \"kubernetes.io/projected/41c0df85-a096-4e29-9dae-4e98007e43ff-kube-api-access-c742l\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.804015 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsw2m\" (UniqueName: \"kubernetes.io/projected/cac2310b-905a-42df-92d8-611fa944c8f6-kube-api-access-tsw2m\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.804023 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41c0df85-a096-4e29-9dae-4e98007e43ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.804031 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjjgb\" (UniqueName: \"kubernetes.io/projected/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec-kube-api-access-tjjgb\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.804040 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/867de14d-11f1-4306-873d-cdc3f6f90216-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.805099 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df74ef9c-2afd-4236-901e-82c48594f70f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "df74ef9c-2afd-4236-901e-82c48594f70f" (UID: "df74ef9c-2afd-4236-901e-82c48594f70f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.805324 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f" (UID: "97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.805715 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a756447-5ce8-4b33-9075-e3b6100d807c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6a756447-5ce8-4b33-9075-e3b6100d807c" (UID: "6a756447-5ce8-4b33-9075-e3b6100d807c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.805753 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e722893-4a07-44e8-9e5a-3e04ff50dc47-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4e722893-4a07-44e8-9e5a-3e04ff50dc47" (UID: "4e722893-4a07-44e8-9e5a-3e04ff50dc47"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.810259 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-kube-api-access-l44fp" (OuterVolumeSpecName: "kube-api-access-l44fp") pod "97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f" (UID: "97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f"). InnerVolumeSpecName "kube-api-access-l44fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.811987 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e722893-4a07-44e8-9e5a-3e04ff50dc47-kube-api-access-zh27s" (OuterVolumeSpecName: "kube-api-access-zh27s") pod "4e722893-4a07-44e8-9e5a-3e04ff50dc47" (UID: "4e722893-4a07-44e8-9e5a-3e04ff50dc47"). InnerVolumeSpecName "kube-api-access-zh27s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.814031 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a756447-5ce8-4b33-9075-e3b6100d807c-kube-api-access-njm48" (OuterVolumeSpecName: "kube-api-access-njm48") pod "6a756447-5ce8-4b33-9075-e3b6100d807c" (UID: "6a756447-5ce8-4b33-9075-e3b6100d807c"). InnerVolumeSpecName "kube-api-access-njm48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.817701 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df74ef9c-2afd-4236-901e-82c48594f70f-kube-api-access-wfhjr" (OuterVolumeSpecName: "kube-api-access-wfhjr") pod "df74ef9c-2afd-4236-901e-82c48594f70f" (UID: "df74ef9c-2afd-4236-901e-82c48594f70f"). InnerVolumeSpecName "kube-api-access-wfhjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.906647 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l44fp\" (UniqueName: \"kubernetes.io/projected/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-kube-api-access-l44fp\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.906675 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a756447-5ce8-4b33-9075-e3b6100d807c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.906753 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfhjr\" (UniqueName: \"kubernetes.io/projected/df74ef9c-2afd-4236-901e-82c48594f70f-kube-api-access-wfhjr\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.906763 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e722893-4a07-44e8-9e5a-3e04ff50dc47-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.906774 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh27s\" (UniqueName: \"kubernetes.io/projected/4e722893-4a07-44e8-9e5a-3e04ff50dc47-kube-api-access-zh27s\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.906782 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df74ef9c-2afd-4236-901e-82c48594f70f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.906791 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:36 crc kubenswrapper[4848]: I1204 14:11:36.906799 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njm48\" (UniqueName: \"kubernetes.io/projected/6a756447-5ce8-4b33-9075-e3b6100d807c-kube-api-access-njm48\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.487355 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"e9cacd34aa9c87a5863f934b6cd848be5fdda6342cc170db3cca10cc8ff209fd"} Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.489070 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"d2af561d-8819-4175-89c0-4a65a661fdb6","Type":"ContainerStarted","Data":"52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10"} Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.492695 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-dchv8" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.493316 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-021e-account-create-update-g264j" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.493389 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-8pzdg" event={"ID":"cf2c3a14-5cfd-49a1-9337-fb0b61611669","Type":"ContainerStarted","Data":"8283fa960e8727cef37f7a575c8d78f76e0df4c960df766a40c8f256345b71c7"} Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.493437 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cxgrr" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.493471 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s2z5f" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.493498 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fd58-account-create-update-r8csb" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.493528 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2804-account-create-update-mrhhd" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.493558 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-61ea-account-create-update-mtrsm" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.493803 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ldg2g" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.521857 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=16.412277704 podStartE2EDuration="19.521837343s" podCreationTimestamp="2025-12-04 14:11:18 +0000 UTC" firstStartedPulling="2025-12-04 14:11:29.753084439 +0000 UTC m=+1393.695580967" lastFinishedPulling="2025-12-04 14:11:32.862644078 +0000 UTC m=+1396.805140606" observedRunningTime="2025-12-04 14:11:37.511894752 +0000 UTC m=+1401.454391290" watchObservedRunningTime="2025-12-04 14:11:37.521837343 +0000 UTC m=+1401.464333871" Dec 04 14:11:37 crc kubenswrapper[4848]: I1204 14:11:37.590514 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-8pzdg" podStartSLOduration=12.923548307 podStartE2EDuration="19.590493728s" podCreationTimestamp="2025-12-04 14:11:18 +0000 UTC" firstStartedPulling="2025-12-04 14:11:29.653992325 +0000 UTC m=+1393.596488853" lastFinishedPulling="2025-12-04 14:11:36.320937746 +0000 UTC m=+1400.263434274" observedRunningTime="2025-12-04 14:11:37.574500511 +0000 UTC m=+1401.516997049" watchObservedRunningTime="2025-12-04 14:11:37.590493728 +0000 UTC m=+1401.532990256" Dec 04 14:11:40 crc kubenswrapper[4848]: I1204 14:11:40.526745 4848 generic.go:334] "Generic (PLEG): container finished" podID="a9e4a10a-a31e-431d-99e2-d4205440fad6" containerID="525d3f55d89a39b18bfc722d1462ff7c6d9b08a8118d7c1217aeabc268276191" exitCode=0 Dec 04 14:11:40 crc kubenswrapper[4848]: I1204 14:11:40.527039 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a9e4a10a-a31e-431d-99e2-d4205440fad6","Type":"ContainerDied","Data":"525d3f55d89a39b18bfc722d1462ff7c6d9b08a8118d7c1217aeabc268276191"} Dec 04 14:11:43 crc kubenswrapper[4848]: I1204 14:11:43.608855 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"6085f8a12efa8dd888e0d19b24545b5b0d97e14b52f6e8af339901a91a9378a2"} Dec 04 14:11:43 crc kubenswrapper[4848]: I1204 14:11:43.609334 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"0601fb59fd38e0f55836705783ef91e94f1c3b2fb7880580ae74916c0c80ac24"} Dec 04 14:11:43 crc kubenswrapper[4848]: I1204 14:11:43.609434 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"de244855640bc58a13716bd3a73ee5b27be22638a7874871d25026f989d6aeac"} Dec 04 14:11:43 crc kubenswrapper[4848]: I1204 14:11:43.610860 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a9e4a10a-a31e-431d-99e2-d4205440fad6","Type":"ContainerStarted","Data":"2b171e294fd7af9ee90d8576c69cd4f07afe7a9f0b1e2427295362343c21e48e"} Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.314485 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.314549 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.314598 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.315386 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"31d6948afbef92b12725ad2b8756f5336adda03c65024c1864ef7ff8a0de5eb8"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.315463 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://31d6948afbef92b12725ad2b8756f5336adda03c65024c1864ef7ff8a0de5eb8" gracePeriod=600 Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.627163 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"fe1c5ef5f53505eadec8fc64f7c74a15a6b3fbaf1312b766d505ddd01a8d81e4"} Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766091 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bscld"] Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766689 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a756447-5ce8-4b33-9075-e3b6100d807c" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766707 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a756447-5ce8-4b33-9075-e3b6100d807c" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766724 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c0df85-a096-4e29-9dae-4e98007e43ff" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766731 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c0df85-a096-4e29-9dae-4e98007e43ff" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766743 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565e9c18-f099-49a4-9457-725b85e0cb7a" containerName="dnsmasq-dns" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766750 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="565e9c18-f099-49a4-9457-725b85e0cb7a" containerName="dnsmasq-dns" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766766 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766772 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766784 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df74ef9c-2afd-4236-901e-82c48594f70f" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766789 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="df74ef9c-2afd-4236-901e-82c48594f70f" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766801 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cac2310b-905a-42df-92d8-611fa944c8f6" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766808 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="cac2310b-905a-42df-92d8-611fa944c8f6" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766815 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c6c058-c395-40c9-b6f1-b1a0580447b5" containerName="ovn-config" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766821 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c6c058-c395-40c9-b6f1-b1a0580447b5" containerName="ovn-config" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766840 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e722893-4a07-44e8-9e5a-3e04ff50dc47" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766846 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e722893-4a07-44e8-9e5a-3e04ff50dc47" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766854 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766860 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766872 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="867de14d-11f1-4306-873d-cdc3f6f90216" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766878 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="867de14d-11f1-4306-873d-cdc3f6f90216" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: E1204 14:11:44.766890 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565e9c18-f099-49a4-9457-725b85e0cb7a" containerName="init" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.766896 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="565e9c18-f099-49a4-9457-725b85e0cb7a" containerName="init" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767081 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767095 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="565e9c18-f099-49a4-9457-725b85e0cb7a" containerName="dnsmasq-dns" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767112 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767121 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="41c0df85-a096-4e29-9dae-4e98007e43ff" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767134 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a756447-5ce8-4b33-9075-e3b6100d807c" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767140 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="df74ef9c-2afd-4236-901e-82c48594f70f" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767148 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="867de14d-11f1-4306-873d-cdc3f6f90216" containerName="mariadb-account-create-update" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767158 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c6c058-c395-40c9-b6f1-b1a0580447b5" containerName="ovn-config" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767166 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e722893-4a07-44e8-9e5a-3e04ff50dc47" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.767177 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="cac2310b-905a-42df-92d8-611fa944c8f6" containerName="mariadb-database-create" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.768514 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.801860 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bscld"] Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.869414 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46j7d\" (UniqueName: \"kubernetes.io/projected/8583821f-0883-4b0c-a021-e1b15d90e249-kube-api-access-46j7d\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.869465 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-utilities\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.869499 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-catalog-content\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.970893 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46j7d\" (UniqueName: \"kubernetes.io/projected/8583821f-0883-4b0c-a021-e1b15d90e249-kube-api-access-46j7d\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.970963 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-utilities\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.970999 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-catalog-content\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.971485 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-catalog-content\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:44 crc kubenswrapper[4848]: I1204 14:11:44.972138 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-utilities\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:45 crc kubenswrapper[4848]: I1204 14:11:45.044351 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46j7d\" (UniqueName: \"kubernetes.io/projected/8583821f-0883-4b0c-a021-e1b15d90e249-kube-api-access-46j7d\") pod \"redhat-marketplace-bscld\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:45 crc kubenswrapper[4848]: I1204 14:11:45.100316 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:45 crc kubenswrapper[4848]: I1204 14:11:45.646659 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bscld"] Dec 04 14:11:45 crc kubenswrapper[4848]: I1204 14:11:45.661754 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="31d6948afbef92b12725ad2b8756f5336adda03c65024c1864ef7ff8a0de5eb8" exitCode=0 Dec 04 14:11:45 crc kubenswrapper[4848]: I1204 14:11:45.663194 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"31d6948afbef92b12725ad2b8756f5336adda03c65024c1864ef7ff8a0de5eb8"} Dec 04 14:11:45 crc kubenswrapper[4848]: I1204 14:11:45.664135 4848 scope.go:117] "RemoveContainer" containerID="131123771f065baa008e65d389e63e335b142b4c147179bfa07ab2fb29d26eae" Dec 04 14:11:45 crc kubenswrapper[4848]: I1204 14:11:45.664416 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357"} Dec 04 14:11:45 crc kubenswrapper[4848]: W1204 14:11:45.848821 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8583821f_0883_4b0c_a021_e1b15d90e249.slice/crio-3573540e0198cb2db7f587d2e33ac3318f1aef28489f1ce4825359920ea58eab WatchSource:0}: Error finding container 3573540e0198cb2db7f587d2e33ac3318f1aef28489f1ce4825359920ea58eab: Status 404 returned error can't find the container with id 3573540e0198cb2db7f587d2e33ac3318f1aef28489f1ce4825359920ea58eab Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.674137 4848 generic.go:334] "Generic (PLEG): container finished" podID="8583821f-0883-4b0c-a021-e1b15d90e249" containerID="5e70c12db217fc6fe35925aab13914cabf950d67d96b20313a7715c04c59b812" exitCode=0 Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.674804 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bscld" event={"ID":"8583821f-0883-4b0c-a021-e1b15d90e249","Type":"ContainerDied","Data":"5e70c12db217fc6fe35925aab13914cabf950d67d96b20313a7715c04c59b812"} Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.674841 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bscld" event={"ID":"8583821f-0883-4b0c-a021-e1b15d90e249","Type":"ContainerStarted","Data":"3573540e0198cb2db7f587d2e33ac3318f1aef28489f1ce4825359920ea58eab"} Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.691041 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"3839c002db9f7b08e42ecad58ff3d6ed063175d4a1c5416811745c20cc9bc9f2"} Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.691078 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"ff4f182f513335110b294824ffcb6c62106648ff8c0e8f5220394cf12e3c7911"} Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.695421 4848 generic.go:334] "Generic (PLEG): container finished" podID="cf2c3a14-5cfd-49a1-9337-fb0b61611669" containerID="8283fa960e8727cef37f7a575c8d78f76e0df4c960df766a40c8f256345b71c7" exitCode=0 Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.695480 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-8pzdg" event={"ID":"cf2c3a14-5cfd-49a1-9337-fb0b61611669","Type":"ContainerDied","Data":"8283fa960e8727cef37f7a575c8d78f76e0df4c960df766a40c8f256345b71c7"} Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.698021 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a9e4a10a-a31e-431d-99e2-d4205440fad6","Type":"ContainerStarted","Data":"c8ca249baff9241ab007ac5a5f98fcc447a4c1566cc3e445c2567650f0331819"} Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.698047 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a9e4a10a-a31e-431d-99e2-d4205440fad6","Type":"ContainerStarted","Data":"6f174ac75267cdf5b65728564c08dd060a9695bf6290b4e1c496b1f9f724a0ef"} Dec 04 14:11:46 crc kubenswrapper[4848]: I1204 14:11:46.734239 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=26.734222022 podStartE2EDuration="26.734222022s" podCreationTimestamp="2025-12-04 14:11:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:46.732382367 +0000 UTC m=+1410.674878895" watchObservedRunningTime="2025-12-04 14:11:46.734222022 +0000 UTC m=+1410.676718550" Dec 04 14:11:47 crc kubenswrapper[4848]: I1204 14:11:47.712245 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bscld" event={"ID":"8583821f-0883-4b0c-a021-e1b15d90e249","Type":"ContainerStarted","Data":"a200ad85ff0920cad63aef5e5f8e860d97b61af6a38b9dd4598241899e92e032"} Dec 04 14:11:47 crc kubenswrapper[4848]: I1204 14:11:47.715899 4848 generic.go:334] "Generic (PLEG): container finished" podID="9925991d-d748-4a39-96a0-3bf2e391cad7" containerID="2ded7473b5d4b3564f6fc47d069bdb2ecb6bd7939ef27e3dfe27f9848e205a6a" exitCode=0 Dec 04 14:11:47 crc kubenswrapper[4848]: I1204 14:11:47.716013 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r7w8t" event={"ID":"9925991d-d748-4a39-96a0-3bf2e391cad7","Type":"ContainerDied","Data":"2ded7473b5d4b3564f6fc47d069bdb2ecb6bd7939ef27e3dfe27f9848e205a6a"} Dec 04 14:11:47 crc kubenswrapper[4848]: I1204 14:11:47.722001 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"9f70e7fce0a41081dd2da5555d49534ecfe2dd6003be5b4bbaa95156a8ed3b40"} Dec 04 14:11:47 crc kubenswrapper[4848]: I1204 14:11:47.722051 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"181543426b1db413bd3a705f7f96d0fda22f1b9663f4e2feab240289c6a10aec"} Dec 04 14:11:47 crc kubenswrapper[4848]: I1204 14:11:47.722063 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"20b2396558d9db0f9e14ea0b870ef0097619cd3dc94c8cbc221616ce07aa567c"} Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.206385 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.343309 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-config-data\") pod \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.343541 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tqnx\" (UniqueName: \"kubernetes.io/projected/cf2c3a14-5cfd-49a1-9337-fb0b61611669-kube-api-access-9tqnx\") pod \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.343572 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-combined-ca-bundle\") pod \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\" (UID: \"cf2c3a14-5cfd-49a1-9337-fb0b61611669\") " Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.350372 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf2c3a14-5cfd-49a1-9337-fb0b61611669-kube-api-access-9tqnx" (OuterVolumeSpecName: "kube-api-access-9tqnx") pod "cf2c3a14-5cfd-49a1-9337-fb0b61611669" (UID: "cf2c3a14-5cfd-49a1-9337-fb0b61611669"). InnerVolumeSpecName "kube-api-access-9tqnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.382981 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf2c3a14-5cfd-49a1-9337-fb0b61611669" (UID: "cf2c3a14-5cfd-49a1-9337-fb0b61611669"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.423212 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-config-data" (OuterVolumeSpecName: "config-data") pod "cf2c3a14-5cfd-49a1-9337-fb0b61611669" (UID: "cf2c3a14-5cfd-49a1-9337-fb0b61611669"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.448574 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.448707 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tqnx\" (UniqueName: \"kubernetes.io/projected/cf2c3a14-5cfd-49a1-9337-fb0b61611669-kube-api-access-9tqnx\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.448758 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf2c3a14-5cfd-49a1-9337-fb0b61611669-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.748551 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"4da9b98f6c7d7b6eece123ce1a10a6f703766c034e6a1dc3b5826821068d37a9"} Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.749150 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0102871d-c950-46e7-af43-a2b3f1cbc39f","Type":"ContainerStarted","Data":"f1c7b6725a6837799f5f2fe0c18d5ac4cc513f1c677d31d2ead6c4f01a78d574"} Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.761880 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-8pzdg" event={"ID":"cf2c3a14-5cfd-49a1-9337-fb0b61611669","Type":"ContainerDied","Data":"cbedabd2535519f9c4c659914407f702e88db180f49b973ff04fbcc9b7f98053"} Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.761944 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbedabd2535519f9c4c659914407f702e88db180f49b973ff04fbcc9b7f98053" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.762001 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-8pzdg" Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.767394 4848 generic.go:334] "Generic (PLEG): container finished" podID="8583821f-0883-4b0c-a021-e1b15d90e249" containerID="a200ad85ff0920cad63aef5e5f8e860d97b61af6a38b9dd4598241899e92e032" exitCode=0 Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.767601 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bscld" event={"ID":"8583821f-0883-4b0c-a021-e1b15d90e249","Type":"ContainerDied","Data":"a200ad85ff0920cad63aef5e5f8e860d97b61af6a38b9dd4598241899e92e032"} Dec 04 14:11:48 crc kubenswrapper[4848]: I1204 14:11:48.816590 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.887355144 podStartE2EDuration="1m5.816573107s" podCreationTimestamp="2025-12-04 14:10:43 +0000 UTC" firstStartedPulling="2025-12-04 14:11:16.999175154 +0000 UTC m=+1380.941671682" lastFinishedPulling="2025-12-04 14:11:45.928393117 +0000 UTC m=+1409.870889645" observedRunningTime="2025-12-04 14:11:48.798812996 +0000 UTC m=+1412.741309534" watchObservedRunningTime="2025-12-04 14:11:48.816573107 +0000 UTC m=+1412.759069635" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.033478 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-wvjbm"] Dec 04 14:11:49 crc kubenswrapper[4848]: E1204 14:11:49.034016 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf2c3a14-5cfd-49a1-9337-fb0b61611669" containerName="keystone-db-sync" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.034043 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf2c3a14-5cfd-49a1-9337-fb0b61611669" containerName="keystone-db-sync" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.035386 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf2c3a14-5cfd-49a1-9337-fb0b61611669" containerName="keystone-db-sync" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.036500 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.055769 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jgxbd"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.058538 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.064457 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4527c" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.064773 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.064893 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.065051 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.065450 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.067160 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.067250 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.067311 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjpqq\" (UniqueName: \"kubernetes.io/projected/1dbb6f5f-8bef-47fc-a203-21536513de20-kube-api-access-mjpqq\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.067331 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-dns-svc\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.067358 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-config\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.070628 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jgxbd"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.124427 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-wvjbm"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.168744 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-combined-ca-bundle\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.168800 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.168868 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-credential-keys\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.168890 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.168907 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-config-data\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.168923 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-fernet-keys\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.168989 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjpqq\" (UniqueName: \"kubernetes.io/projected/1dbb6f5f-8bef-47fc-a203-21536513de20-kube-api-access-mjpqq\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.169015 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-dns-svc\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.169044 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-config\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.169080 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hhnz\" (UniqueName: \"kubernetes.io/projected/f69e1612-8e36-464d-a216-7a13ea62dbd0-kube-api-access-9hhnz\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.169115 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-scripts\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.169914 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-sb\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.170506 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-nb\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.171440 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-dns-svc\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.171940 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-config\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.203011 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-ggd6n"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.210734 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.231253 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.231942 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-bhc6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.232804 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-ggd6n"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.238878 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjpqq\" (UniqueName: \"kubernetes.io/projected/1dbb6f5f-8bef-47fc-a203-21536513de20-kube-api-access-mjpqq\") pod \"dnsmasq-dns-f877ddd87-wvjbm\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.272336 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-combined-ca-bundle\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.272653 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hhnz\" (UniqueName: \"kubernetes.io/projected/f69e1612-8e36-464d-a216-7a13ea62dbd0-kube-api-access-9hhnz\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.273181 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p997\" (UniqueName: \"kubernetes.io/projected/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-kube-api-access-5p997\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.273307 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-scripts\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.277413 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-config-data\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.277539 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-combined-ca-bundle\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.277708 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-credential-keys\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.277793 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-config-data\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.277868 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-fernet-keys\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.285083 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-fernet-keys\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.289800 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-combined-ca-bundle\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.300865 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-credential-keys\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.301600 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-scripts\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.311832 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-config-data\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.351556 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.352411 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hhnz\" (UniqueName: \"kubernetes.io/projected/f69e1612-8e36-464d-a216-7a13ea62dbd0-kube-api-access-9hhnz\") pod \"keystone-bootstrap-jgxbd\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.377883 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-6hkhx"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.383361 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p997\" (UniqueName: \"kubernetes.io/projected/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-kube-api-access-5p997\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.383493 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-config-data\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.383725 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-combined-ca-bundle\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.391279 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.395785 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-config-data\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.396300 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-combined-ca-bundle\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.411011 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.427236 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tltfn" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.434928 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.435308 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.436269 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p997\" (UniqueName: \"kubernetes.io/projected/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-kube-api-access-5p997\") pod \"heat-db-sync-ggd6n\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.489541 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5ls8\" (UniqueName: \"kubernetes.io/projected/f64ca832-b3c0-45a9-b967-a8f3ce9af156-kube-api-access-m5ls8\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.489704 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-combined-ca-bundle\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.489805 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-config\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.504242 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-4dpj6"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.508897 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.544734 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-qdffl" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.545671 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.545894 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.575105 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ggd6n" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.592464 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-db-sync-config-data\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.597897 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4dpj6"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.617747 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-etc-machine-id\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.618003 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5ls8\" (UniqueName: \"kubernetes.io/projected/f64ca832-b3c0-45a9-b967-a8f3ce9af156-kube-api-access-m5ls8\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.621469 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-combined-ca-bundle\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.621722 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzhzf\" (UniqueName: \"kubernetes.io/projected/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-kube-api-access-lzhzf\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.621889 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-scripts\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.632106 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-combined-ca-bundle\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.712062 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-config\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.712503 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-combined-ca-bundle\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.712790 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-config-data\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.736311 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5ls8\" (UniqueName: \"kubernetes.io/projected/f64ca832-b3c0-45a9-b967-a8f3ce9af156-kube-api-access-m5ls8\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.783475 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6hkhx"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.805891 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-config\") pod \"neutron-db-sync-6hkhx\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.824224 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-combined-ca-bundle\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.828484 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-config-data\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.828659 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-db-sync-config-data\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.828781 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-etc-machine-id\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.828937 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzhzf\" (UniqueName: \"kubernetes.io/projected/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-kube-api-access-lzhzf\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.829042 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-scripts\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.830158 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-etc-machine-id\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.834238 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-combined-ca-bundle\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.846063 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-config-data\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.849383 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-scripts\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.849677 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-db-sync-config-data\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.873283 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-wvjbm"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.882199 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzhzf\" (UniqueName: \"kubernetes.io/projected/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-kube-api-access-lzhzf\") pod \"cinder-db-sync-4dpj6\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.922911 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.974062 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-l5jbz"] Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.976367 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.982567 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-q6vwc" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.982798 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 14:11:49 crc kubenswrapper[4848]: I1204 14:11:49.998810 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-bsgwh"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.016210 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.024668 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.034583 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.066242 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-l5jbz"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.169695 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.178123 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.178257 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpl44\" (UniqueName: \"kubernetes.io/projected/23eae77d-63ca-4004-969b-cfac163d3975-kube-api-access-xpl44\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.178287 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-combined-ca-bundle\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.184061 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-svc\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.184218 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-config\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.184341 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq57b\" (UniqueName: \"kubernetes.io/projected/ea545629-3f98-4f0e-97a1-17eae9485b65-kube-api-access-fq57b\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.184431 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-db-sync-config-data\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.184535 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.286449 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-bsgwh"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.292614 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-db-sync-config-data\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.293740 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.296029 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.296291 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.294853 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.296430 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpl44\" (UniqueName: \"kubernetes.io/projected/23eae77d-63ca-4004-969b-cfac163d3975-kube-api-access-xpl44\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.296632 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-combined-ca-bundle\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.296903 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-svc\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.297125 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-config\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.297324 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq57b\" (UniqueName: \"kubernetes.io/projected/ea545629-3f98-4f0e-97a1-17eae9485b65-kube-api-access-fq57b\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.298422 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-svc\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.298764 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.299579 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-db-sync-config-data\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.300862 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.301110 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-config\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.305559 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-combined-ca-bundle\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.318523 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-m9h4t"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.320205 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.325985 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.330932 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq57b\" (UniqueName: \"kubernetes.io/projected/ea545629-3f98-4f0e-97a1-17eae9485b65-kube-api-access-fq57b\") pod \"dnsmasq-dns-5959f8865f-bsgwh\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.331091 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.331189 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-k5j2j" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.333550 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpl44\" (UniqueName: \"kubernetes.io/projected/23eae77d-63ca-4004-969b-cfac163d3975-kube-api-access-xpl44\") pod \"barbican-db-sync-l5jbz\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.338272 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.348754 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-m9h4t"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.375777 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-bsgwh"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.376777 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.402166 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-combined-ca-bundle\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.402784 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-config-data\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.402901 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-scripts\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.403261 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7380f023-2712-4a2b-a193-ec401d07c9ad-logs\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.403349 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmkjn\" (UniqueName: \"kubernetes.io/projected/7380f023-2712-4a2b-a193-ec401d07c9ad-kube-api-access-vmkjn\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.415493 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bscld" podStartSLOduration=3.9024959470000002 podStartE2EDuration="6.415470027s" podCreationTimestamp="2025-12-04 14:11:44 +0000 UTC" firstStartedPulling="2025-12-04 14:11:46.676148753 +0000 UTC m=+1410.618645281" lastFinishedPulling="2025-12-04 14:11:49.189122833 +0000 UTC m=+1413.131619361" observedRunningTime="2025-12-04 14:11:49.944992076 +0000 UTC m=+1413.887488604" watchObservedRunningTime="2025-12-04 14:11:50.415470027 +0000 UTC m=+1414.357966555" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.462776 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c2csj"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.466312 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c2csj"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.466582 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.469611 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.503272 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.508396 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.508612 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.512764 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.514127 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7380f023-2712-4a2b-a193-ec401d07c9ad-logs\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.517896 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.520461 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7380f023-2712-4a2b-a193-ec401d07c9ad-logs\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.521429 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmkjn\" (UniqueName: \"kubernetes.io/projected/7380f023-2712-4a2b-a193-ec401d07c9ad-kube-api-access-vmkjn\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.521554 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.521850 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-config\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.522029 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.526272 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-combined-ca-bundle\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.526539 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp8wf\" (UniqueName: \"kubernetes.io/projected/39da2201-07e4-4c24-9c73-3fb261551a77-kube-api-access-hp8wf\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.527264 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.527416 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-config-data\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.527549 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-scripts\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.539977 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-config-data\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.564383 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-combined-ca-bundle\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.564682 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-scripts\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.568385 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmkjn\" (UniqueName: \"kubernetes.io/projected/7380f023-2712-4a2b-a193-ec401d07c9ad-kube-api-access-vmkjn\") pod \"placement-db-sync-m9h4t\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.602415 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.631741 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.631794 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-config\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.631821 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.631848 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632010 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp8wf\" (UniqueName: \"kubernetes.io/projected/39da2201-07e4-4c24-9c73-3fb261551a77-kube-api-access-hp8wf\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632035 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-config-data\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632075 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-log-httpd\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632111 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-scripts\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632153 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmzt6\" (UniqueName: \"kubernetes.io/projected/99e49830-7634-4257-af28-938f373af926-kube-api-access-zmzt6\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632180 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632200 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-run-httpd\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632226 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.632265 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.634426 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.636588 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-config\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.637240 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.637728 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.639935 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.665086 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-m9h4t" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.681118 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp8wf\" (UniqueName: \"kubernetes.io/projected/39da2201-07e4-4c24-9c73-3fb261551a77-kube-api-access-hp8wf\") pod \"dnsmasq-dns-58dd9ff6bc-c2csj\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.685082 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.687471 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-wvjbm"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.733146 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw2gh\" (UniqueName: \"kubernetes.io/projected/9925991d-d748-4a39-96a0-3bf2e391cad7-kube-api-access-qw2gh\") pod \"9925991d-d748-4a39-96a0-3bf2e391cad7\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.733222 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-combined-ca-bundle\") pod \"9925991d-d748-4a39-96a0-3bf2e391cad7\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.733347 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-db-sync-config-data\") pod \"9925991d-d748-4a39-96a0-3bf2e391cad7\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.733461 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-config-data\") pod \"9925991d-d748-4a39-96a0-3bf2e391cad7\" (UID: \"9925991d-d748-4a39-96a0-3bf2e391cad7\") " Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.733941 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.734016 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-config-data\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.734041 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-log-httpd\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.734073 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-scripts\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.734112 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmzt6\" (UniqueName: \"kubernetes.io/projected/99e49830-7634-4257-af28-938f373af926-kube-api-access-zmzt6\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.734160 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-run-httpd\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.734196 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.736714 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9925991d-d748-4a39-96a0-3bf2e391cad7-kube-api-access-qw2gh" (OuterVolumeSpecName: "kube-api-access-qw2gh") pod "9925991d-d748-4a39-96a0-3bf2e391cad7" (UID: "9925991d-d748-4a39-96a0-3bf2e391cad7"). InnerVolumeSpecName "kube-api-access-qw2gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.741674 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-config-data\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.741863 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.742194 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-log-httpd\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.742589 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-run-httpd\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.745824 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-scripts\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.746346 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9925991d-d748-4a39-96a0-3bf2e391cad7" (UID: "9925991d-d748-4a39-96a0-3bf2e391cad7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.751408 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.761889 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmzt6\" (UniqueName: \"kubernetes.io/projected/99e49830-7634-4257-af28-938f373af926-kube-api-access-zmzt6\") pod \"ceilometer-0\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " pod="openstack/ceilometer-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.791333 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9925991d-d748-4a39-96a0-3bf2e391cad7" (UID: "9925991d-d748-4a39-96a0-3bf2e391cad7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.832318 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-config-data" (OuterVolumeSpecName: "config-data") pod "9925991d-d748-4a39-96a0-3bf2e391cad7" (UID: "9925991d-d748-4a39-96a0-3bf2e391cad7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.836903 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.836939 4848 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.836960 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9925991d-d748-4a39-96a0-3bf2e391cad7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.836968 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw2gh\" (UniqueName: \"kubernetes.io/projected/9925991d-d748-4a39-96a0-3bf2e391cad7-kube-api-access-qw2gh\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.854652 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jgxbd"] Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.855442 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.855458 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:50 crc kubenswrapper[4848]: W1204 14:11:50.868544 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf69e1612_8e36_464d_a216_7a13ea62dbd0.slice/crio-df1be2e728c468056b9639b53909419ff1eb3324918a2943569e330a36170b66 WatchSource:0}: Error finding container df1be2e728c468056b9639b53909419ff1eb3324918a2943569e330a36170b66: Status 404 returned error can't find the container with id df1be2e728c468056b9639b53909419ff1eb3324918a2943569e330a36170b66 Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.870914 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-ggd6n"] Dec 04 14:11:50 crc kubenswrapper[4848]: W1204 14:11:50.871425 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba8b0d2_85d2_45c6_9612_0652e5dfdd33.slice/crio-54ca3357c9f824eb53198308f66403f243da82dbacb0262d1cf73f921777696d WatchSource:0}: Error finding container 54ca3357c9f824eb53198308f66403f243da82dbacb0262d1cf73f921777696d: Status 404 returned error can't find the container with id 54ca3357c9f824eb53198308f66403f243da82dbacb0262d1cf73f921777696d Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.914201 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:50 crc kubenswrapper[4848]: I1204 14:11:50.983652 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bscld" event={"ID":"8583821f-0883-4b0c-a021-e1b15d90e249","Type":"ContainerStarted","Data":"be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d"} Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.004754 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r7w8t" event={"ID":"9925991d-d748-4a39-96a0-3bf2e391cad7","Type":"ContainerDied","Data":"50a37c81c8ab5d5bedaa655a73650b51d6e7b4d285da531ee9d8cf62b0ffae67"} Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.005679 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50a37c81c8ab5d5bedaa655a73650b51d6e7b4d285da531ee9d8cf62b0ffae67" Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.005367 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r7w8t" Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.016545 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ggd6n" event={"ID":"dba8b0d2-85d2-45c6-9612-0652e5dfdd33","Type":"ContainerStarted","Data":"54ca3357c9f824eb53198308f66403f243da82dbacb0262d1cf73f921777696d"} Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.017939 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.026147 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jgxbd" event={"ID":"f69e1612-8e36-464d-a216-7a13ea62dbd0","Type":"ContainerStarted","Data":"df1be2e728c468056b9639b53909419ff1eb3324918a2943569e330a36170b66"} Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.047123 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" event={"ID":"1dbb6f5f-8bef-47fc-a203-21536513de20","Type":"ContainerStarted","Data":"e3ac6c784d6eb8921b7c99f65439b0707d4b2ee1767d71de862d7502fc4417d2"} Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.082250 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.119470 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4dpj6"] Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.456623 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-6hkhx"] Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.481270 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-bsgwh"] Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.506956 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-l5jbz"] Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.835249 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c2csj"] Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.889557 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-m9h4t"] Dec 04 14:11:51 crc kubenswrapper[4848]: I1204 14:11:51.994252 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c2csj"] Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.094240 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-sx2nq"] Dec 04 14:11:52 crc kubenswrapper[4848]: E1204 14:11:52.094702 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9925991d-d748-4a39-96a0-3bf2e391cad7" containerName="glance-db-sync" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.094713 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9925991d-d748-4a39-96a0-3bf2e391cad7" containerName="glance-db-sync" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.094935 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9925991d-d748-4a39-96a0-3bf2e391cad7" containerName="glance-db-sync" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.096009 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.116528 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" event={"ID":"ea545629-3f98-4f0e-97a1-17eae9485b65","Type":"ContainerStarted","Data":"eaa0b5c5cedcc9f186cb2ceedfbfb0f39c0a118ba304a0ab2220e3dd96193521"} Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.130076 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-sx2nq"] Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.154851 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" event={"ID":"39da2201-07e4-4c24-9c73-3fb261551a77","Type":"ContainerStarted","Data":"350245042a0db34b0dc671d2617da765151bfda6160f17d260247948938338e0"} Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.168615 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6hkhx" event={"ID":"f64ca832-b3c0-45a9-b967-a8f3ce9af156","Type":"ContainerStarted","Data":"15a28dd8f0b2bc39c2246d8377669117344254ecbff73f329c55abe00c4b573e"} Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.170727 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jgxbd" event={"ID":"f69e1612-8e36-464d-a216-7a13ea62dbd0","Type":"ContainerStarted","Data":"34efa360e1a91ddb73a5293b1621957a6d5eaba3f5b9101dedb9b126e90776d5"} Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.225702 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.230128 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.230272 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.230325 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-config\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.230366 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv6w5\" (UniqueName: \"kubernetes.io/projected/4e1bd61d-621d-4c6a-84d6-5981897a8320-kube-api-access-kv6w5\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.230438 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.249391 4848 generic.go:334] "Generic (PLEG): container finished" podID="1dbb6f5f-8bef-47fc-a203-21536513de20" containerID="115c2aa27f3d8bc300cf05f77176351c29aaa2c7151a24c3dc36b51706a7910b" exitCode=0 Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.249502 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" event={"ID":"1dbb6f5f-8bef-47fc-a203-21536513de20","Type":"ContainerDied","Data":"115c2aa27f3d8bc300cf05f77176351c29aaa2c7151a24c3dc36b51706a7910b"} Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.272307 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-m9h4t" event={"ID":"7380f023-2712-4a2b-a193-ec401d07c9ad","Type":"ContainerStarted","Data":"bdbe8f1defca2fbb457b9f3e5971657fef8967088fed05531c5409659b0aaf41"} Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.301404 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-l5jbz" event={"ID":"23eae77d-63ca-4004-969b-cfac163d3975","Type":"ContainerStarted","Data":"a81e9d511512ada4ab85d1225e9043ed8750085cee32fb0001d67566e0fd29ca"} Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.338892 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.338992 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.339047 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.339075 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-config\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.339103 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv6w5\" (UniqueName: \"kubernetes.io/projected/4e1bd61d-621d-4c6a-84d6-5981897a8320-kube-api-access-kv6w5\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.339138 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.339417 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4dpj6" event={"ID":"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5","Type":"ContainerStarted","Data":"1c1c180a6983b04cf02e03a08cf1b43c4584ffacd439497a5667613e5c029dc7"} Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.340024 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.340606 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-config\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.346582 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.347364 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.360200 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.364924 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.384770 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv6w5\" (UniqueName: \"kubernetes.io/projected/4e1bd61d-621d-4c6a-84d6-5981897a8320-kube-api-access-kv6w5\") pod \"dnsmasq-dns-785d8bcb8c-sx2nq\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.389739 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jgxbd" podStartSLOduration=3.389700171 podStartE2EDuration="3.389700171s" podCreationTimestamp="2025-12-04 14:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:52.325597056 +0000 UTC m=+1416.268093584" watchObservedRunningTime="2025-12-04 14:11:52.389700171 +0000 UTC m=+1416.332196699" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.497146 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.519010 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.605034 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.616736 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.644848 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.655192 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rzzqm" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.655977 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.732022 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.795352 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-scripts\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.795401 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.795480 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-logs\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.795501 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.795573 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-config-data\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.795599 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.795672 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chjbq\" (UniqueName: \"kubernetes.io/projected/5f3842b6-52cb-49a4-9001-e9dd954a3b16-kube-api-access-chjbq\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.921559 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-logs\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.921617 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.921813 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-config-data\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.921855 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.922069 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chjbq\" (UniqueName: \"kubernetes.io/projected/5f3842b6-52cb-49a4-9001-e9dd954a3b16-kube-api-access-chjbq\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.922139 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-scripts\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.922162 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.922596 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.922820 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-logs\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.923064 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.934580 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-config-data\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.934719 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-scripts\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.956349 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:52 crc kubenswrapper[4848]: I1204 14:11:52.988841 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chjbq\" (UniqueName: \"kubernetes.io/projected/5f3842b6-52cb-49a4-9001-e9dd954a3b16-kube-api-access-chjbq\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.099590 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " pod="openstack/glance-default-external-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.151030 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.238702 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-config\") pod \"1dbb6f5f-8bef-47fc-a203-21536513de20\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.239245 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-nb\") pod \"1dbb6f5f-8bef-47fc-a203-21536513de20\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.239300 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjpqq\" (UniqueName: \"kubernetes.io/projected/1dbb6f5f-8bef-47fc-a203-21536513de20-kube-api-access-mjpqq\") pod \"1dbb6f5f-8bef-47fc-a203-21536513de20\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.239331 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-sb\") pod \"1dbb6f5f-8bef-47fc-a203-21536513de20\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.239382 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-dns-svc\") pod \"1dbb6f5f-8bef-47fc-a203-21536513de20\" (UID: \"1dbb6f5f-8bef-47fc-a203-21536513de20\") " Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.253770 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dbb6f5f-8bef-47fc-a203-21536513de20-kube-api-access-mjpqq" (OuterVolumeSpecName: "kube-api-access-mjpqq") pod "1dbb6f5f-8bef-47fc-a203-21536513de20" (UID: "1dbb6f5f-8bef-47fc-a203-21536513de20"). InnerVolumeSpecName "kube-api-access-mjpqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.269120 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1dbb6f5f-8bef-47fc-a203-21536513de20" (UID: "1dbb6f5f-8bef-47fc-a203-21536513de20"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.272500 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1dbb6f5f-8bef-47fc-a203-21536513de20" (UID: "1dbb6f5f-8bef-47fc-a203-21536513de20"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.294984 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1dbb6f5f-8bef-47fc-a203-21536513de20" (UID: "1dbb6f5f-8bef-47fc-a203-21536513de20"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.319365 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-config" (OuterVolumeSpecName: "config") pod "1dbb6f5f-8bef-47fc-a203-21536513de20" (UID: "1dbb6f5f-8bef-47fc-a203-21536513de20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.342906 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.342933 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.342946 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjpqq\" (UniqueName: \"kubernetes.io/projected/1dbb6f5f-8bef-47fc-a203-21536513de20-kube-api-access-mjpqq\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.342967 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.342975 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1dbb6f5f-8bef-47fc-a203-21536513de20-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.348926 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.417019 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99e49830-7634-4257-af28-938f373af926","Type":"ContainerStarted","Data":"9dea439f054cfaf9d67c7b11625a8be33e60f1e0416b83f1defc9451ff6186b4"} Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.422848 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" event={"ID":"1dbb6f5f-8bef-47fc-a203-21536513de20","Type":"ContainerDied","Data":"e3ac6c784d6eb8921b7c99f65439b0707d4b2ee1767d71de862d7502fc4417d2"} Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.422903 4848 scope.go:117] "RemoveContainer" containerID="115c2aa27f3d8bc300cf05f77176351c29aaa2c7151a24c3dc36b51706a7910b" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.423098 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f877ddd87-wvjbm" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.446367 4848 generic.go:334] "Generic (PLEG): container finished" podID="ea545629-3f98-4f0e-97a1-17eae9485b65" containerID="be4590cbeed62d5ae89694adcda42ea0f4766db6b56bf172d6ad0f1d38524e16" exitCode=0 Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.446684 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" event={"ID":"ea545629-3f98-4f0e-97a1-17eae9485b65","Type":"ContainerDied","Data":"be4590cbeed62d5ae89694adcda42ea0f4766db6b56bf172d6ad0f1d38524e16"} Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.449634 4848 generic.go:334] "Generic (PLEG): container finished" podID="39da2201-07e4-4c24-9c73-3fb261551a77" containerID="1cfa8bfd71969c84684e50b9dcc3af103331da8145092ac4da3be57ff145747e" exitCode=0 Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.449899 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" event={"ID":"39da2201-07e4-4c24-9c73-3fb261551a77","Type":"ContainerDied","Data":"1cfa8bfd71969c84684e50b9dcc3af103331da8145092ac4da3be57ff145747e"} Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.484074 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6hkhx" event={"ID":"f64ca832-b3c0-45a9-b967-a8f3ce9af156","Type":"ContainerStarted","Data":"12c88f7d42b296bb37e58011445ef99d0f1697bb57702df574b9329bcbdf2f39"} Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.544177 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-6hkhx" podStartSLOduration=4.54415057 podStartE2EDuration="4.54415057s" podCreationTimestamp="2025-12-04 14:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:53.518503089 +0000 UTC m=+1417.460999617" watchObservedRunningTime="2025-12-04 14:11:53.54415057 +0000 UTC m=+1417.486647098" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.686447 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-wvjbm"] Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.702833 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f877ddd87-wvjbm"] Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.712941 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:11:53 crc kubenswrapper[4848]: E1204 14:11:53.713651 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dbb6f5f-8bef-47fc-a203-21536513de20" containerName="init" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.713705 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dbb6f5f-8bef-47fc-a203-21536513de20" containerName="init" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.714166 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dbb6f5f-8bef-47fc-a203-21536513de20" containerName="init" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.717069 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.721379 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.738700 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.770321 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-sx2nq"] Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.878766 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.878884 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.878924 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.878965 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgk84\" (UniqueName: \"kubernetes.io/projected/a5134303-3212-4d92-a897-bb3af9324743-kube-api-access-vgk84\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.878998 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.879013 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.879262 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.982612 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.982943 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.983007 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgk84\" (UniqueName: \"kubernetes.io/projected/a5134303-3212-4d92-a897-bb3af9324743-kube-api-access-vgk84\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.983076 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.983092 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.983218 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.983310 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.985898 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.986209 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.986261 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.989795 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:53 crc kubenswrapper[4848]: I1204 14:11:53.990758 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.006910 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgk84\" (UniqueName: \"kubernetes.io/projected/a5134303-3212-4d92-a897-bb3af9324743-kube-api-access-vgk84\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.009438 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.057351 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.113427 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.136289 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.299507 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-config\") pod \"ea545629-3f98-4f0e-97a1-17eae9485b65\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.299871 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-sb\") pod \"ea545629-3f98-4f0e-97a1-17eae9485b65\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.299978 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-svc\") pod \"ea545629-3f98-4f0e-97a1-17eae9485b65\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.300053 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq57b\" (UniqueName: \"kubernetes.io/projected/ea545629-3f98-4f0e-97a1-17eae9485b65-kube-api-access-fq57b\") pod \"ea545629-3f98-4f0e-97a1-17eae9485b65\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.300070 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-swift-storage-0\") pod \"ea545629-3f98-4f0e-97a1-17eae9485b65\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.300172 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-nb\") pod \"ea545629-3f98-4f0e-97a1-17eae9485b65\" (UID: \"ea545629-3f98-4f0e-97a1-17eae9485b65\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.315531 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea545629-3f98-4f0e-97a1-17eae9485b65-kube-api-access-fq57b" (OuterVolumeSpecName: "kube-api-access-fq57b") pod "ea545629-3f98-4f0e-97a1-17eae9485b65" (UID: "ea545629-3f98-4f0e-97a1-17eae9485b65"). InnerVolumeSpecName "kube-api-access-fq57b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.318634 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.358410 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-config" (OuterVolumeSpecName: "config") pod "ea545629-3f98-4f0e-97a1-17eae9485b65" (UID: "ea545629-3f98-4f0e-97a1-17eae9485b65"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.374362 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea545629-3f98-4f0e-97a1-17eae9485b65" (UID: "ea545629-3f98-4f0e-97a1-17eae9485b65"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.406645 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea545629-3f98-4f0e-97a1-17eae9485b65" (UID: "ea545629-3f98-4f0e-97a1-17eae9485b65"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.411235 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.411263 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.411273 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.411282 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq57b\" (UniqueName: \"kubernetes.io/projected/ea545629-3f98-4f0e-97a1-17eae9485b65-kube-api-access-fq57b\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.438793 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ea545629-3f98-4f0e-97a1-17eae9485b65" (UID: "ea545629-3f98-4f0e-97a1-17eae9485b65"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.440628 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea545629-3f98-4f0e-97a1-17eae9485b65" (UID: "ea545629-3f98-4f0e-97a1-17eae9485b65"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.513825 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.513855 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea545629-3f98-4f0e-97a1-17eae9485b65-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.522540 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dbb6f5f-8bef-47fc-a203-21536513de20" path="/var/lib/kubelet/pods/1dbb6f5f-8bef-47fc-a203-21536513de20/volumes" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.609600 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" event={"ID":"ea545629-3f98-4f0e-97a1-17eae9485b65","Type":"ContainerDied","Data":"eaa0b5c5cedcc9f186cb2ceedfbfb0f39c0a118ba304a0ab2220e3dd96193521"} Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.609656 4848 scope.go:117] "RemoveContainer" containerID="be4590cbeed62d5ae89694adcda42ea0f4766db6b56bf172d6ad0f1d38524e16" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.609771 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-bsgwh" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.615689 4848 generic.go:334] "Generic (PLEG): container finished" podID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerID="e0e35ed00f7715ec74f7a9d7407871e7d9c4652d57aa8a2ad307a9676d943187" exitCode=0 Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.615751 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" event={"ID":"4e1bd61d-621d-4c6a-84d6-5981897a8320","Type":"ContainerDied","Data":"e0e35ed00f7715ec74f7a9d7407871e7d9c4652d57aa8a2ad307a9676d943187"} Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.615785 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" event={"ID":"4e1bd61d-621d-4c6a-84d6-5981897a8320","Type":"ContainerStarted","Data":"9bf205f797a4aefb95d3d570284c9cc88faaafe11ffb93d2287667740b114358"} Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.622218 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f3842b6-52cb-49a4-9001-e9dd954a3b16","Type":"ContainerStarted","Data":"7bb8fc40bf92b488876b384bca59749f2db76e51e167605d5e52afc3ab49cd33"} Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.727039 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.758437 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.822779 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-config\") pod \"39da2201-07e4-4c24-9c73-3fb261551a77\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.822853 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-nb\") pod \"39da2201-07e4-4c24-9c73-3fb261551a77\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.822892 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-swift-storage-0\") pod \"39da2201-07e4-4c24-9c73-3fb261551a77\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.823015 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp8wf\" (UniqueName: \"kubernetes.io/projected/39da2201-07e4-4c24-9c73-3fb261551a77-kube-api-access-hp8wf\") pod \"39da2201-07e4-4c24-9c73-3fb261551a77\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.823062 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-svc\") pod \"39da2201-07e4-4c24-9c73-3fb261551a77\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.823184 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-sb\") pod \"39da2201-07e4-4c24-9c73-3fb261551a77\" (UID: \"39da2201-07e4-4c24-9c73-3fb261551a77\") " Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.834014 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39da2201-07e4-4c24-9c73-3fb261551a77-kube-api-access-hp8wf" (OuterVolumeSpecName: "kube-api-access-hp8wf") pod "39da2201-07e4-4c24-9c73-3fb261551a77" (UID: "39da2201-07e4-4c24-9c73-3fb261551a77"). InnerVolumeSpecName "kube-api-access-hp8wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.845073 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-bsgwh"] Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.880960 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-bsgwh"] Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.882832 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "39da2201-07e4-4c24-9c73-3fb261551a77" (UID: "39da2201-07e4-4c24-9c73-3fb261551a77"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.883042 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "39da2201-07e4-4c24-9c73-3fb261551a77" (UID: "39da2201-07e4-4c24-9c73-3fb261551a77"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.898632 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39da2201-07e4-4c24-9c73-3fb261551a77" (UID: "39da2201-07e4-4c24-9c73-3fb261551a77"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.902038 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "39da2201-07e4-4c24-9c73-3fb261551a77" (UID: "39da2201-07e4-4c24-9c73-3fb261551a77"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.911592 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-config" (OuterVolumeSpecName: "config") pod "39da2201-07e4-4c24-9c73-3fb261551a77" (UID: "39da2201-07e4-4c24-9c73-3fb261551a77"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.926907 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp8wf\" (UniqueName: \"kubernetes.io/projected/39da2201-07e4-4c24-9c73-3fb261551a77-kube-api-access-hp8wf\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.926978 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.926995 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.927007 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.927019 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:54 crc kubenswrapper[4848]: I1204 14:11:54.927030 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39da2201-07e4-4c24-9c73-3fb261551a77-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.100543 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.103479 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.650445 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5134303-3212-4d92-a897-bb3af9324743","Type":"ContainerStarted","Data":"b432259f1567806c1b16fb9810ceaf786a592eb5038c2edd1fa4f9298e02672d"} Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.677456 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" event={"ID":"4e1bd61d-621d-4c6a-84d6-5981897a8320","Type":"ContainerStarted","Data":"40c77f9f36a9e6dee29394145e40a1fd23c2d22573461a8ab118f9ad085233d3"} Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.678346 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.713769 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.716269 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-c2csj" event={"ID":"39da2201-07e4-4c24-9c73-3fb261551a77","Type":"ContainerDied","Data":"350245042a0db34b0dc671d2617da765151bfda6160f17d260247948938338e0"} Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.716323 4848 scope.go:117] "RemoveContainer" containerID="1cfa8bfd71969c84684e50b9dcc3af103331da8145092ac4da3be57ff145747e" Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.786562 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" podStartSLOduration=4.786538108 podStartE2EDuration="4.786538108s" podCreationTimestamp="2025-12-04 14:11:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:55.72478596 +0000 UTC m=+1419.667282498" watchObservedRunningTime="2025-12-04 14:11:55.786538108 +0000 UTC m=+1419.729034636" Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.835668 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c2csj"] Dec 04 14:11:55 crc kubenswrapper[4848]: I1204 14:11:55.846346 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-c2csj"] Dec 04 14:11:56 crc kubenswrapper[4848]: I1204 14:11:56.233796 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-bscld" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="registry-server" probeResult="failure" output=< Dec 04 14:11:56 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 14:11:56 crc kubenswrapper[4848]: > Dec 04 14:11:56 crc kubenswrapper[4848]: I1204 14:11:56.411429 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39da2201-07e4-4c24-9c73-3fb261551a77" path="/var/lib/kubelet/pods/39da2201-07e4-4c24-9c73-3fb261551a77/volumes" Dec 04 14:11:56 crc kubenswrapper[4848]: I1204 14:11:56.412302 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea545629-3f98-4f0e-97a1-17eae9485b65" path="/var/lib/kubelet/pods/ea545629-3f98-4f0e-97a1-17eae9485b65/volumes" Dec 04 14:11:56 crc kubenswrapper[4848]: I1204 14:11:56.735148 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5134303-3212-4d92-a897-bb3af9324743","Type":"ContainerStarted","Data":"fd4c38891c836464e661fb989f36fa41a992ca8b0a805d1a2fcf3dcb3d59bd1a"} Dec 04 14:11:56 crc kubenswrapper[4848]: I1204 14:11:56.737957 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f3842b6-52cb-49a4-9001-e9dd954a3b16","Type":"ContainerStarted","Data":"d848e21a3fe0aec03dfcfc9b130ef03915d6fdabc6ea37d576ed9fccec41600c"} Dec 04 14:11:57 crc kubenswrapper[4848]: I1204 14:11:57.782360 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5134303-3212-4d92-a897-bb3af9324743","Type":"ContainerStarted","Data":"7969ec3ac31829d04ce9339340264b03221f04a18fb147604361d59c21114e2e"} Dec 04 14:11:57 crc kubenswrapper[4848]: I1204 14:11:57.852774 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.852747142 podStartE2EDuration="5.852747142s" podCreationTimestamp="2025-12-04 14:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:57.841284845 +0000 UTC m=+1421.783781383" watchObservedRunningTime="2025-12-04 14:11:57.852747142 +0000 UTC m=+1421.795243670" Dec 04 14:11:58 crc kubenswrapper[4848]: I1204 14:11:58.801777 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f3842b6-52cb-49a4-9001-e9dd954a3b16","Type":"ContainerStarted","Data":"b9129b9c914805f556a273a20348b843ca0906c1b25cda045e56bcad43416698"} Dec 04 14:11:58 crc kubenswrapper[4848]: I1204 14:11:58.832664 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.832638429 podStartE2EDuration="6.832638429s" podCreationTimestamp="2025-12-04 14:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:58.820909754 +0000 UTC m=+1422.763406282" watchObservedRunningTime="2025-12-04 14:11:58.832638429 +0000 UTC m=+1422.775134957" Dec 04 14:11:59 crc kubenswrapper[4848]: I1204 14:11:59.817209 4848 generic.go:334] "Generic (PLEG): container finished" podID="f69e1612-8e36-464d-a216-7a13ea62dbd0" containerID="34efa360e1a91ddb73a5293b1621957a6d5eaba3f5b9101dedb9b126e90776d5" exitCode=0 Dec 04 14:11:59 crc kubenswrapper[4848]: I1204 14:11:59.817304 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jgxbd" event={"ID":"f69e1612-8e36-464d-a216-7a13ea62dbd0","Type":"ContainerDied","Data":"34efa360e1a91ddb73a5293b1621957a6d5eaba3f5b9101dedb9b126e90776d5"} Dec 04 14:11:59 crc kubenswrapper[4848]: I1204 14:11:59.833406 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:11:59 crc kubenswrapper[4848]: I1204 14:11:59.917912 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:11:59 crc kubenswrapper[4848]: I1204 14:11:59.918341 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a5134303-3212-4d92-a897-bb3af9324743" containerName="glance-log" containerID="cri-o://fd4c38891c836464e661fb989f36fa41a992ca8b0a805d1a2fcf3dcb3d59bd1a" gracePeriod=30 Dec 04 14:11:59 crc kubenswrapper[4848]: I1204 14:11:59.918454 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a5134303-3212-4d92-a897-bb3af9324743" containerName="glance-httpd" containerID="cri-o://7969ec3ac31829d04ce9339340264b03221f04a18fb147604361d59c21114e2e" gracePeriod=30 Dec 04 14:12:00 crc kubenswrapper[4848]: I1204 14:12:00.848834 4848 generic.go:334] "Generic (PLEG): container finished" podID="a5134303-3212-4d92-a897-bb3af9324743" containerID="7969ec3ac31829d04ce9339340264b03221f04a18fb147604361d59c21114e2e" exitCode=0 Dec 04 14:12:00 crc kubenswrapper[4848]: I1204 14:12:00.848906 4848 generic.go:334] "Generic (PLEG): container finished" podID="a5134303-3212-4d92-a897-bb3af9324743" containerID="fd4c38891c836464e661fb989f36fa41a992ca8b0a805d1a2fcf3dcb3d59bd1a" exitCode=143 Dec 04 14:12:00 crc kubenswrapper[4848]: I1204 14:12:00.848912 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5134303-3212-4d92-a897-bb3af9324743","Type":"ContainerDied","Data":"7969ec3ac31829d04ce9339340264b03221f04a18fb147604361d59c21114e2e"} Dec 04 14:12:00 crc kubenswrapper[4848]: I1204 14:12:00.848986 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5134303-3212-4d92-a897-bb3af9324743","Type":"ContainerDied","Data":"fd4c38891c836464e661fb989f36fa41a992ca8b0a805d1a2fcf3dcb3d59bd1a"} Dec 04 14:12:00 crc kubenswrapper[4848]: I1204 14:12:00.849458 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerName="glance-httpd" containerID="cri-o://b9129b9c914805f556a273a20348b843ca0906c1b25cda045e56bcad43416698" gracePeriod=30 Dec 04 14:12:00 crc kubenswrapper[4848]: I1204 14:12:00.850295 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerName="glance-log" containerID="cri-o://d848e21a3fe0aec03dfcfc9b130ef03915d6fdabc6ea37d576ed9fccec41600c" gracePeriod=30 Dec 04 14:12:01 crc kubenswrapper[4848]: I1204 14:12:01.893657 4848 generic.go:334] "Generic (PLEG): container finished" podID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerID="b9129b9c914805f556a273a20348b843ca0906c1b25cda045e56bcad43416698" exitCode=0 Dec 04 14:12:01 crc kubenswrapper[4848]: I1204 14:12:01.894208 4848 generic.go:334] "Generic (PLEG): container finished" podID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerID="d848e21a3fe0aec03dfcfc9b130ef03915d6fdabc6ea37d576ed9fccec41600c" exitCode=143 Dec 04 14:12:01 crc kubenswrapper[4848]: I1204 14:12:01.894232 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f3842b6-52cb-49a4-9001-e9dd954a3b16","Type":"ContainerDied","Data":"b9129b9c914805f556a273a20348b843ca0906c1b25cda045e56bcad43416698"} Dec 04 14:12:01 crc kubenswrapper[4848]: I1204 14:12:01.894290 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f3842b6-52cb-49a4-9001-e9dd954a3b16","Type":"ContainerDied","Data":"d848e21a3fe0aec03dfcfc9b130ef03915d6fdabc6ea37d576ed9fccec41600c"} Dec 04 14:12:02 crc kubenswrapper[4848]: I1204 14:12:02.498780 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:12:02 crc kubenswrapper[4848]: I1204 14:12:02.577444 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-x9k96"] Dec 04 14:12:02 crc kubenswrapper[4848]: I1204 14:12:02.577856 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-x9k96" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" containerID="cri-o://1e568a7ab9c8b37d95568cc59c1b6e0a0a7d7328b565bd47cb8c987ae3780f0d" gracePeriod=10 Dec 04 14:12:03 crc kubenswrapper[4848]: I1204 14:12:03.951961 4848 generic.go:334] "Generic (PLEG): container finished" podID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerID="1e568a7ab9c8b37d95568cc59c1b6e0a0a7d7328b565bd47cb8c987ae3780f0d" exitCode=0 Dec 04 14:12:03 crc kubenswrapper[4848]: I1204 14:12:03.952217 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-x9k96" event={"ID":"c8675f3d-03ac-46e5-be9e-84991aa3e80a","Type":"ContainerDied","Data":"1e568a7ab9c8b37d95568cc59c1b6e0a0a7d7328b565bd47cb8c987ae3780f0d"} Dec 04 14:12:04 crc kubenswrapper[4848]: I1204 14:12:04.639808 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-x9k96" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: connect: connection refused" Dec 04 14:12:05 crc kubenswrapper[4848]: I1204 14:12:05.145348 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:12:05 crc kubenswrapper[4848]: I1204 14:12:05.206072 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:12:05 crc kubenswrapper[4848]: I1204 14:12:05.381599 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bscld"] Dec 04 14:12:06 crc kubenswrapper[4848]: I1204 14:12:06.993783 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bscld" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="registry-server" containerID="cri-o://be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" gracePeriod=2 Dec 04 14:12:08 crc kubenswrapper[4848]: E1204 14:12:08.892511 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 04 14:12:08 crc kubenswrapper[4848]: E1204 14:12:08.893221 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n569h65dh86h57fh666h64fh678h7dh5cbh5c4h574h5bbh99h5c5h5fh6h5bh647hcchd7hf9h55hfbh6fh57dh66h67fh5chch664h549h678q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zmzt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(99e49830-7634-4257-af28-938f373af926): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:12:09 crc kubenswrapper[4848]: I1204 14:12:09.023532 4848 generic.go:334] "Generic (PLEG): container finished" podID="8583821f-0883-4b0c-a021-e1b15d90e249" containerID="be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" exitCode=0 Dec 04 14:12:09 crc kubenswrapper[4848]: I1204 14:12:09.023599 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bscld" event={"ID":"8583821f-0883-4b0c-a021-e1b15d90e249","Type":"ContainerDied","Data":"be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d"} Dec 04 14:12:09 crc kubenswrapper[4848]: E1204 14:12:09.467871 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 04 14:12:09 crc kubenswrapper[4848]: E1204 14:12:09.468121 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xpl44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-l5jbz_openstack(23eae77d-63ca-4004-969b-cfac163d3975): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:12:09 crc kubenswrapper[4848]: E1204 14:12:09.469412 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-l5jbz" podUID="23eae77d-63ca-4004-969b-cfac163d3975" Dec 04 14:12:09 crc kubenswrapper[4848]: I1204 14:12:09.639231 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-x9k96" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: connect: connection refused" Dec 04 14:12:10 crc kubenswrapper[4848]: E1204 14:12:10.035691 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-l5jbz" podUID="23eae77d-63ca-4004-969b-cfac163d3975" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.554763 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.594156 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-combined-ca-bundle\") pod \"f69e1612-8e36-464d-a216-7a13ea62dbd0\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.594208 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-config-data\") pod \"f69e1612-8e36-464d-a216-7a13ea62dbd0\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.594235 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-fernet-keys\") pod \"f69e1612-8e36-464d-a216-7a13ea62dbd0\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.594359 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-scripts\") pod \"f69e1612-8e36-464d-a216-7a13ea62dbd0\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.594384 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hhnz\" (UniqueName: \"kubernetes.io/projected/f69e1612-8e36-464d-a216-7a13ea62dbd0-kube-api-access-9hhnz\") pod \"f69e1612-8e36-464d-a216-7a13ea62dbd0\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.594415 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-credential-keys\") pod \"f69e1612-8e36-464d-a216-7a13ea62dbd0\" (UID: \"f69e1612-8e36-464d-a216-7a13ea62dbd0\") " Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.613710 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69e1612-8e36-464d-a216-7a13ea62dbd0-kube-api-access-9hhnz" (OuterVolumeSpecName: "kube-api-access-9hhnz") pod "f69e1612-8e36-464d-a216-7a13ea62dbd0" (UID: "f69e1612-8e36-464d-a216-7a13ea62dbd0"). InnerVolumeSpecName "kube-api-access-9hhnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.616465 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-scripts" (OuterVolumeSpecName: "scripts") pod "f69e1612-8e36-464d-a216-7a13ea62dbd0" (UID: "f69e1612-8e36-464d-a216-7a13ea62dbd0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.620367 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f69e1612-8e36-464d-a216-7a13ea62dbd0" (UID: "f69e1612-8e36-464d-a216-7a13ea62dbd0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.626496 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f69e1612-8e36-464d-a216-7a13ea62dbd0" (UID: "f69e1612-8e36-464d-a216-7a13ea62dbd0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.634652 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f69e1612-8e36-464d-a216-7a13ea62dbd0" (UID: "f69e1612-8e36-464d-a216-7a13ea62dbd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.642397 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-x9k96" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: connect: connection refused" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.642525 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.649380 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-config-data" (OuterVolumeSpecName: "config-data") pod "f69e1612-8e36-464d-a216-7a13ea62dbd0" (UID: "f69e1612-8e36-464d-a216-7a13ea62dbd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.698158 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.698200 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.698211 4848 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.698224 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.698237 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hhnz\" (UniqueName: \"kubernetes.io/projected/f69e1612-8e36-464d-a216-7a13ea62dbd0-kube-api-access-9hhnz\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:14 crc kubenswrapper[4848]: I1204 14:12:14.698249 4848 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f69e1612-8e36-464d-a216-7a13ea62dbd0-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.085058 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jgxbd" event={"ID":"f69e1612-8e36-464d-a216-7a13ea62dbd0","Type":"ContainerDied","Data":"df1be2e728c468056b9639b53909419ff1eb3324918a2943569e330a36170b66"} Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.085109 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df1be2e728c468056b9639b53909419ff1eb3324918a2943569e330a36170b66" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.085171 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jgxbd" Dec 04 14:12:15 crc kubenswrapper[4848]: E1204 14:12:15.101723 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d is running failed: container process not found" containerID="be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:12:15 crc kubenswrapper[4848]: E1204 14:12:15.102177 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d is running failed: container process not found" containerID="be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:12:15 crc kubenswrapper[4848]: E1204 14:12:15.102478 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d is running failed: container process not found" containerID="be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:12:15 crc kubenswrapper[4848]: E1204 14:12:15.102510 4848 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-bscld" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="registry-server" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.755076 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jgxbd"] Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.760906 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jgxbd"] Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.850623 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-t9r82"] Dec 04 14:12:15 crc kubenswrapper[4848]: E1204 14:12:15.851064 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea545629-3f98-4f0e-97a1-17eae9485b65" containerName="init" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.851078 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea545629-3f98-4f0e-97a1-17eae9485b65" containerName="init" Dec 04 14:12:15 crc kubenswrapper[4848]: E1204 14:12:15.851095 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39da2201-07e4-4c24-9c73-3fb261551a77" containerName="init" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.851100 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="39da2201-07e4-4c24-9c73-3fb261551a77" containerName="init" Dec 04 14:12:15 crc kubenswrapper[4848]: E1204 14:12:15.851114 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69e1612-8e36-464d-a216-7a13ea62dbd0" containerName="keystone-bootstrap" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.851121 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69e1612-8e36-464d-a216-7a13ea62dbd0" containerName="keystone-bootstrap" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.851325 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69e1612-8e36-464d-a216-7a13ea62dbd0" containerName="keystone-bootstrap" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.851351 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="39da2201-07e4-4c24-9c73-3fb261551a77" containerName="init" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.851366 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea545629-3f98-4f0e-97a1-17eae9485b65" containerName="init" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.852153 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.854117 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.854580 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.856229 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4527c" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.856244 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.856238 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.863494 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t9r82"] Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.937381 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-config-data\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.937426 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-credential-keys\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.937494 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-scripts\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.937563 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-combined-ca-bundle\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.937590 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzt65\" (UniqueName: \"kubernetes.io/projected/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-kube-api-access-jzt65\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:15 crc kubenswrapper[4848]: I1204 14:12:15.937619 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-fernet-keys\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.039246 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-combined-ca-bundle\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.039503 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzt65\" (UniqueName: \"kubernetes.io/projected/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-kube-api-access-jzt65\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.039535 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-fernet-keys\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.040426 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-config-data\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.040451 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-credential-keys\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.040471 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-scripts\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.045114 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-credential-keys\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.045474 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-fernet-keys\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.045617 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-config-data\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.045702 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-combined-ca-bundle\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.046013 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-scripts\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.054731 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzt65\" (UniqueName: \"kubernetes.io/projected/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-kube-api-access-jzt65\") pod \"keystone-bootstrap-t9r82\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.190208 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:16 crc kubenswrapper[4848]: E1204 14:12:16.315858 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 04 14:12:16 crc kubenswrapper[4848]: E1204 14:12:16.316315 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vmkjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-m9h4t_openstack(7380f023-2712-4a2b-a193-ec401d07c9ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:12:16 crc kubenswrapper[4848]: E1204 14:12:16.318341 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-m9h4t" podUID="7380f023-2712-4a2b-a193-ec401d07c9ad" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.410925 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f69e1612-8e36-464d-a216-7a13ea62dbd0" path="/var/lib/kubelet/pods/f69e1612-8e36-464d-a216-7a13ea62dbd0/volumes" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.421149 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.451146 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-config-data\") pod \"a5134303-3212-4d92-a897-bb3af9324743\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.451186 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"a5134303-3212-4d92-a897-bb3af9324743\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.452567 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-httpd-run\") pod \"a5134303-3212-4d92-a897-bb3af9324743\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.452682 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-logs\") pod \"a5134303-3212-4d92-a897-bb3af9324743\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.452713 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-scripts\") pod \"a5134303-3212-4d92-a897-bb3af9324743\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.452746 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-combined-ca-bundle\") pod \"a5134303-3212-4d92-a897-bb3af9324743\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.452831 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgk84\" (UniqueName: \"kubernetes.io/projected/a5134303-3212-4d92-a897-bb3af9324743-kube-api-access-vgk84\") pod \"a5134303-3212-4d92-a897-bb3af9324743\" (UID: \"a5134303-3212-4d92-a897-bb3af9324743\") " Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.455691 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a5134303-3212-4d92-a897-bb3af9324743" (UID: "a5134303-3212-4d92-a897-bb3af9324743"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.456393 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-logs" (OuterVolumeSpecName: "logs") pod "a5134303-3212-4d92-a897-bb3af9324743" (UID: "a5134303-3212-4d92-a897-bb3af9324743"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.478778 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "a5134303-3212-4d92-a897-bb3af9324743" (UID: "a5134303-3212-4d92-a897-bb3af9324743"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.478815 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5134303-3212-4d92-a897-bb3af9324743-kube-api-access-vgk84" (OuterVolumeSpecName: "kube-api-access-vgk84") pod "a5134303-3212-4d92-a897-bb3af9324743" (UID: "a5134303-3212-4d92-a897-bb3af9324743"). InnerVolumeSpecName "kube-api-access-vgk84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.480175 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-scripts" (OuterVolumeSpecName: "scripts") pod "a5134303-3212-4d92-a897-bb3af9324743" (UID: "a5134303-3212-4d92-a897-bb3af9324743"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.512602 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5134303-3212-4d92-a897-bb3af9324743" (UID: "a5134303-3212-4d92-a897-bb3af9324743"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.544139 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-config-data" (OuterVolumeSpecName: "config-data") pod "a5134303-3212-4d92-a897-bb3af9324743" (UID: "a5134303-3212-4d92-a897-bb3af9324743"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.555736 4848 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.555771 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5134303-3212-4d92-a897-bb3af9324743-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.555779 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.555788 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.555803 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgk84\" (UniqueName: \"kubernetes.io/projected/a5134303-3212-4d92-a897-bb3af9324743-kube-api-access-vgk84\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.555811 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5134303-3212-4d92-a897-bb3af9324743-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.555849 4848 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.582498 4848 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 04 14:12:16 crc kubenswrapper[4848]: I1204 14:12:16.657601 4848 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.111077 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5134303-3212-4d92-a897-bb3af9324743","Type":"ContainerDied","Data":"b432259f1567806c1b16fb9810ceaf786a592eb5038c2edd1fa4f9298e02672d"} Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.112383 4848 scope.go:117] "RemoveContainer" containerID="7969ec3ac31829d04ce9339340264b03221f04a18fb147604361d59c21114e2e" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.111135 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: E1204 14:12:17.113199 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-m9h4t" podUID="7380f023-2712-4a2b-a193-ec401d07c9ad" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.167051 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.216291 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.229489 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:12:17 crc kubenswrapper[4848]: E1204 14:12:17.230031 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5134303-3212-4d92-a897-bb3af9324743" containerName="glance-httpd" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.230051 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5134303-3212-4d92-a897-bb3af9324743" containerName="glance-httpd" Dec 04 14:12:17 crc kubenswrapper[4848]: E1204 14:12:17.230082 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5134303-3212-4d92-a897-bb3af9324743" containerName="glance-log" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.230089 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5134303-3212-4d92-a897-bb3af9324743" containerName="glance-log" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.230280 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5134303-3212-4d92-a897-bb3af9324743" containerName="glance-httpd" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.230303 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5134303-3212-4d92-a897-bb3af9324743" containerName="glance-log" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.231585 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.245592 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.245918 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.251101 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.275351 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.275465 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.275520 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.275570 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-logs\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.275622 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.275646 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.275669 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2gcv\" (UniqueName: \"kubernetes.io/projected/f252353f-4af7-445f-9860-30d4299d8ad6-kube-api-access-j2gcv\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.275757 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379291 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379370 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-logs\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379415 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379431 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379448 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2gcv\" (UniqueName: \"kubernetes.io/projected/f252353f-4af7-445f-9860-30d4299d8ad6-kube-api-access-j2gcv\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379512 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379589 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379646 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.379872 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.380145 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.380195 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-logs\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.384904 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.386159 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.387388 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.387861 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.396233 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2gcv\" (UniqueName: \"kubernetes.io/projected/f252353f-4af7-445f-9860-30d4299d8ad6-kube-api-access-j2gcv\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.437229 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:12:17 crc kubenswrapper[4848]: I1204 14:12:17.565773 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:18 crc kubenswrapper[4848]: I1204 14:12:18.407137 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5134303-3212-4d92-a897-bb3af9324743" path="/var/lib/kubelet/pods/a5134303-3212-4d92-a897-bb3af9324743/volumes" Dec 04 14:12:22 crc kubenswrapper[4848]: I1204 14:12:22.163645 4848 generic.go:334] "Generic (PLEG): container finished" podID="f64ca832-b3c0-45a9-b967-a8f3ce9af156" containerID="12c88f7d42b296bb37e58011445ef99d0f1697bb57702df574b9329bcbdf2f39" exitCode=0 Dec 04 14:12:22 crc kubenswrapper[4848]: I1204 14:12:22.163711 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6hkhx" event={"ID":"f64ca832-b3c0-45a9-b967-a8f3ce9af156","Type":"ContainerDied","Data":"12c88f7d42b296bb37e58011445ef99d0f1697bb57702df574b9329bcbdf2f39"} Dec 04 14:12:23 crc kubenswrapper[4848]: I1204 14:12:23.349966 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:12:23 crc kubenswrapper[4848]: I1204 14:12:23.350244 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:12:24 crc kubenswrapper[4848]: I1204 14:12:24.639561 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-x9k96" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: i/o timeout" Dec 04 14:12:25 crc kubenswrapper[4848]: E1204 14:12:25.100977 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d is running failed: container process not found" containerID="be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:12:25 crc kubenswrapper[4848]: E1204 14:12:25.101815 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d is running failed: container process not found" containerID="be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:12:25 crc kubenswrapper[4848]: E1204 14:12:25.102454 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d is running failed: container process not found" containerID="be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" cmd=["grpc_health_probe","-addr=:50051"] Dec 04 14:12:25 crc kubenswrapper[4848]: E1204 14:12:25.102494 4848 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-bscld" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="registry-server" Dec 04 14:12:25 crc kubenswrapper[4848]: E1204 14:12:25.146711 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 04 14:12:25 crc kubenswrapper[4848]: E1204 14:12:25.146850 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5p997,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-ggd6n_openstack(dba8b0d2-85d2-45c6-9612-0652e5dfdd33): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:12:25 crc kubenswrapper[4848]: E1204 14:12:25.148147 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-ggd6n" podUID="dba8b0d2-85d2-45c6-9612-0652e5dfdd33" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.213655 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-6hkhx" event={"ID":"f64ca832-b3c0-45a9-b967-a8f3ce9af156","Type":"ContainerDied","Data":"15a28dd8f0b2bc39c2246d8377669117344254ecbff73f329c55abe00c4b573e"} Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.213706 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15a28dd8f0b2bc39c2246d8377669117344254ecbff73f329c55abe00c4b573e" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.227171 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-x9k96" event={"ID":"c8675f3d-03ac-46e5-be9e-84991aa3e80a","Type":"ContainerDied","Data":"7b7f0648a88dcaf1406f9f211e6ffacc6ee471c71d70b23ba7108b2e92bc6cf8"} Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.227225 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b7f0648a88dcaf1406f9f211e6ffacc6ee471c71d70b23ba7108b2e92bc6cf8" Dec 04 14:12:25 crc kubenswrapper[4848]: E1204 14:12:25.228629 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-ggd6n" podUID="dba8b0d2-85d2-45c6-9612-0652e5dfdd33" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.351403 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.364753 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.371363 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.381561 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.502789 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46j7d\" (UniqueName: \"kubernetes.io/projected/8583821f-0883-4b0c-a021-e1b15d90e249-kube-api-access-46j7d\") pod \"8583821f-0883-4b0c-a021-e1b15d90e249\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.502846 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-config-data\") pod \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.502916 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-nb\") pod \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.502944 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-config\") pod \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503034 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chjbq\" (UniqueName: \"kubernetes.io/projected/5f3842b6-52cb-49a4-9001-e9dd954a3b16-kube-api-access-chjbq\") pod \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503070 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-config\") pod \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503087 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-scripts\") pod \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503110 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-utilities\") pod \"8583821f-0883-4b0c-a021-e1b15d90e249\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503151 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-combined-ca-bundle\") pod \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503177 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-dns-svc\") pod \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503196 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-combined-ca-bundle\") pod \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503220 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-logs\") pod \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503241 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-sb\") pod \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503262 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54grj\" (UniqueName: \"kubernetes.io/projected/c8675f3d-03ac-46e5-be9e-84991aa3e80a-kube-api-access-54grj\") pod \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\" (UID: \"c8675f3d-03ac-46e5-be9e-84991aa3e80a\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503296 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-catalog-content\") pod \"8583821f-0883-4b0c-a021-e1b15d90e249\" (UID: \"8583821f-0883-4b0c-a021-e1b15d90e249\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503335 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5ls8\" (UniqueName: \"kubernetes.io/projected/f64ca832-b3c0-45a9-b967-a8f3ce9af156-kube-api-access-m5ls8\") pod \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\" (UID: \"f64ca832-b3c0-45a9-b967-a8f3ce9af156\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503359 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-httpd-run\") pod \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.503390 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\" (UID: \"5f3842b6-52cb-49a4-9001-e9dd954a3b16\") " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.504572 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-logs" (OuterVolumeSpecName: "logs") pod "5f3842b6-52cb-49a4-9001-e9dd954a3b16" (UID: "5f3842b6-52cb-49a4-9001-e9dd954a3b16"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.504974 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5f3842b6-52cb-49a4-9001-e9dd954a3b16" (UID: "5f3842b6-52cb-49a4-9001-e9dd954a3b16"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.505070 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-utilities" (OuterVolumeSpecName: "utilities") pod "8583821f-0883-4b0c-a021-e1b15d90e249" (UID: "8583821f-0883-4b0c-a021-e1b15d90e249"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.510761 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f64ca832-b3c0-45a9-b967-a8f3ce9af156-kube-api-access-m5ls8" (OuterVolumeSpecName: "kube-api-access-m5ls8") pod "f64ca832-b3c0-45a9-b967-a8f3ce9af156" (UID: "f64ca832-b3c0-45a9-b967-a8f3ce9af156"). InnerVolumeSpecName "kube-api-access-m5ls8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.510673 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "5f3842b6-52cb-49a4-9001-e9dd954a3b16" (UID: "5f3842b6-52cb-49a4-9001-e9dd954a3b16"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.514071 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-scripts" (OuterVolumeSpecName: "scripts") pod "5f3842b6-52cb-49a4-9001-e9dd954a3b16" (UID: "5f3842b6-52cb-49a4-9001-e9dd954a3b16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.515733 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8675f3d-03ac-46e5-be9e-84991aa3e80a-kube-api-access-54grj" (OuterVolumeSpecName: "kube-api-access-54grj") pod "c8675f3d-03ac-46e5-be9e-84991aa3e80a" (UID: "c8675f3d-03ac-46e5-be9e-84991aa3e80a"). InnerVolumeSpecName "kube-api-access-54grj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.518223 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8583821f-0883-4b0c-a021-e1b15d90e249" (UID: "8583821f-0883-4b0c-a021-e1b15d90e249"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.526188 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f3842b6-52cb-49a4-9001-e9dd954a3b16-kube-api-access-chjbq" (OuterVolumeSpecName: "kube-api-access-chjbq") pod "5f3842b6-52cb-49a4-9001-e9dd954a3b16" (UID: "5f3842b6-52cb-49a4-9001-e9dd954a3b16"). InnerVolumeSpecName "kube-api-access-chjbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.543338 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8583821f-0883-4b0c-a021-e1b15d90e249-kube-api-access-46j7d" (OuterVolumeSpecName: "kube-api-access-46j7d") pod "8583821f-0883-4b0c-a021-e1b15d90e249" (UID: "8583821f-0883-4b0c-a021-e1b15d90e249"). InnerVolumeSpecName "kube-api-access-46j7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.548612 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-config" (OuterVolumeSpecName: "config") pod "f64ca832-b3c0-45a9-b967-a8f3ce9af156" (UID: "f64ca832-b3c0-45a9-b967-a8f3ce9af156"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.549437 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f3842b6-52cb-49a4-9001-e9dd954a3b16" (UID: "5f3842b6-52cb-49a4-9001-e9dd954a3b16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.569032 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f64ca832-b3c0-45a9-b967-a8f3ce9af156" (UID: "f64ca832-b3c0-45a9-b967-a8f3ce9af156"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.600744 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-config" (OuterVolumeSpecName: "config") pod "c8675f3d-03ac-46e5-be9e-84991aa3e80a" (UID: "c8675f3d-03ac-46e5-be9e-84991aa3e80a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.602052 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c8675f3d-03ac-46e5-be9e-84991aa3e80a" (UID: "c8675f3d-03ac-46e5-be9e-84991aa3e80a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.602440 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-config-data" (OuterVolumeSpecName: "config-data") pod "5f3842b6-52cb-49a4-9001-e9dd954a3b16" (UID: "5f3842b6-52cb-49a4-9001-e9dd954a3b16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.605762 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c8675f3d-03ac-46e5-be9e-84991aa3e80a" (UID: "c8675f3d-03ac-46e5-be9e-84991aa3e80a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606670 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606695 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606706 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chjbq\" (UniqueName: \"kubernetes.io/projected/5f3842b6-52cb-49a4-9001-e9dd954a3b16-kube-api-access-chjbq\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606719 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606728 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606738 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606748 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64ca832-b3c0-45a9-b967-a8f3ce9af156-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606756 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606764 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606772 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606780 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54grj\" (UniqueName: \"kubernetes.io/projected/c8675f3d-03ac-46e5-be9e-84991aa3e80a-kube-api-access-54grj\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606788 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8583821f-0883-4b0c-a021-e1b15d90e249-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606797 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5ls8\" (UniqueName: \"kubernetes.io/projected/f64ca832-b3c0-45a9-b967-a8f3ce9af156-kube-api-access-m5ls8\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606805 4848 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5f3842b6-52cb-49a4-9001-e9dd954a3b16-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606833 4848 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606843 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46j7d\" (UniqueName: \"kubernetes.io/projected/8583821f-0883-4b0c-a021-e1b15d90e249-kube-api-access-46j7d\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.606852 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3842b6-52cb-49a4-9001-e9dd954a3b16-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.611031 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c8675f3d-03ac-46e5-be9e-84991aa3e80a" (UID: "c8675f3d-03ac-46e5-be9e-84991aa3e80a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.653159 4848 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.707507 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c8675f3d-03ac-46e5-be9e-84991aa3e80a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:25 crc kubenswrapper[4848]: I1204 14:12:25.707546 4848 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.244107 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bscld" event={"ID":"8583821f-0883-4b0c-a021-e1b15d90e249","Type":"ContainerDied","Data":"3573540e0198cb2db7f587d2e33ac3318f1aef28489f1ce4825359920ea58eab"} Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.244406 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bscld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.246856 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-6hkhx" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.247339 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5f3842b6-52cb-49a4-9001-e9dd954a3b16","Type":"ContainerDied","Data":"7bb8fc40bf92b488876b384bca59749f2db76e51e167605d5e52afc3ab49cd33"} Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.247400 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.247815 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-x9k96" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.325059 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bscld"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.341001 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bscld"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.356202 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.371620 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.383901 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-x9k96"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.392717 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-x9k96"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.413798 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" path="/var/lib/kubelet/pods/5f3842b6-52cb-49a4-9001-e9dd954a3b16/volumes" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.415340 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" path="/var/lib/kubelet/pods/8583821f-0883-4b0c-a021-e1b15d90e249/volumes" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.417026 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" path="/var/lib/kubelet/pods/c8675f3d-03ac-46e5-be9e-84991aa3e80a/volumes" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.418848 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:12:26 crc kubenswrapper[4848]: E1204 14:12:26.419643 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerName="glance-log" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.419668 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerName="glance-log" Dec 04 14:12:26 crc kubenswrapper[4848]: E1204 14:12:26.419691 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="registry-server" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.419733 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="registry-server" Dec 04 14:12:26 crc kubenswrapper[4848]: E1204 14:12:26.419752 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="extract-content" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.419761 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="extract-content" Dec 04 14:12:26 crc kubenswrapper[4848]: E1204 14:12:26.419774 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f64ca832-b3c0-45a9-b967-a8f3ce9af156" containerName="neutron-db-sync" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.419783 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f64ca832-b3c0-45a9-b967-a8f3ce9af156" containerName="neutron-db-sync" Dec 04 14:12:26 crc kubenswrapper[4848]: E1204 14:12:26.419803 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.419812 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" Dec 04 14:12:26 crc kubenswrapper[4848]: E1204 14:12:26.419861 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="extract-utilities" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.419871 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="extract-utilities" Dec 04 14:12:26 crc kubenswrapper[4848]: E1204 14:12:26.419886 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="init" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.419894 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="init" Dec 04 14:12:26 crc kubenswrapper[4848]: E1204 14:12:26.420028 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerName="glance-httpd" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.420041 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerName="glance-httpd" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.420407 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerName="glance-log" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.420440 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.420454 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f3842b6-52cb-49a4-9001-e9dd954a3b16" containerName="glance-httpd" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.420482 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f64ca832-b3c0-45a9-b967-a8f3ce9af156" containerName="neutron-db-sync" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.420501 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="8583821f-0883-4b0c-a021-e1b15d90e249" containerName="registry-server" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.424372 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.424450 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.429705 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.430327 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.611198 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kmgld"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.612846 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.626441 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-scripts\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.626545 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.626639 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-logs\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.626703 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cl5d\" (UniqueName: \"kubernetes.io/projected/db70a1ef-eaa8-4da0-8abe-72aa5c822226-kube-api-access-2cl5d\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.626724 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-config-data\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.626740 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.626764 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.626779 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.631673 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kmgld"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.728851 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.728898 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95bpb\" (UniqueName: \"kubernetes.io/projected/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-kube-api-access-95bpb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.728970 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-scripts\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729008 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729045 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-svc\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729068 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729130 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-config\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729158 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-logs\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729206 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729226 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-config-data\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729240 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cl5d\" (UniqueName: \"kubernetes.io/projected/db70a1ef-eaa8-4da0-8abe-72aa5c822226-kube-api-access-2cl5d\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729259 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729279 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.729298 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.731661 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-logs\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.732325 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.734901 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.737694 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-scripts\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.739707 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.752892 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.760710 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-config-data\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.769735 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cl5d\" (UniqueName: \"kubernetes.io/projected/db70a1ef-eaa8-4da0-8abe-72aa5c822226-kube-api-access-2cl5d\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.777241 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cc6cc5676-4j9jn"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.778663 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " pod="openstack/glance-default-external-api-0" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.780450 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.782621 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.783636 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tltfn" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.783777 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.783895 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.789209 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cc6cc5676-4j9jn"] Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.831800 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-config\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.832094 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.832319 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.832346 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95bpb\" (UniqueName: \"kubernetes.io/projected/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-kube-api-access-95bpb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.832433 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.832707 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-svc\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.832980 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.833770 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.833790 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-svc\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.834284 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.834717 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-config\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.854103 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95bpb\" (UniqueName: \"kubernetes.io/projected/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-kube-api-access-95bpb\") pod \"dnsmasq-dns-55f844cf75-kmgld\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.934739 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qht5p\" (UniqueName: \"kubernetes.io/projected/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-kube-api-access-qht5p\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.934777 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-config\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.934796 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-ovndb-tls-certs\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.934814 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-combined-ca-bundle\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.934842 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-httpd-config\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:26 crc kubenswrapper[4848]: I1204 14:12:26.945170 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.036461 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qht5p\" (UniqueName: \"kubernetes.io/projected/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-kube-api-access-qht5p\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.036500 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-config\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.036521 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-ovndb-tls-certs\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.036538 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-combined-ca-bundle\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.036572 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-httpd-config\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.040661 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-config\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.040810 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-httpd-config\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.042466 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-combined-ca-bundle\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.053893 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.055182 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qht5p\" (UniqueName: \"kubernetes.io/projected/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-kube-api-access-qht5p\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.055610 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-ovndb-tls-certs\") pod \"neutron-6cc6cc5676-4j9jn\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.227319 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:27 crc kubenswrapper[4848]: E1204 14:12:27.462087 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 04 14:12:27 crc kubenswrapper[4848]: E1204 14:12:27.462257 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lzhzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-4dpj6_openstack(47da25b0-48bd-4dd1-8ad9-6db3f8565ad5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:12:27 crc kubenswrapper[4848]: E1204 14:12:27.463661 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-4dpj6" podUID="47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" Dec 04 14:12:27 crc kubenswrapper[4848]: I1204 14:12:27.947489 4848 scope.go:117] "RemoveContainer" containerID="fd4c38891c836464e661fb989f36fa41a992ca8b0a805d1a2fcf3dcb3d59bd1a" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.141240 4848 scope.go:117] "RemoveContainer" containerID="be7d77ac0e129000e1c5f41b5a268618c16b7e0ca319ca32f8a9d826213b650d" Dec 04 14:12:28 crc kubenswrapper[4848]: E1204 14:12:28.328872 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-4dpj6" podUID="47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.493807 4848 scope.go:117] "RemoveContainer" containerID="a200ad85ff0920cad63aef5e5f8e860d97b61af6a38b9dd4598241899e92e032" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.550004 4848 scope.go:117] "RemoveContainer" containerID="5e70c12db217fc6fe35925aab13914cabf950d67d96b20313a7715c04c59b812" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.602198 4848 scope.go:117] "RemoveContainer" containerID="b9129b9c914805f556a273a20348b843ca0906c1b25cda045e56bcad43416698" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.627200 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t9r82"] Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.661509 4848 scope.go:117] "RemoveContainer" containerID="d848e21a3fe0aec03dfcfc9b130ef03915d6fdabc6ea37d576ed9fccec41600c" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.665404 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.829037 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:12:28 crc kubenswrapper[4848]: W1204 14:12:28.833530 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf252353f_4af7_445f_9860_30d4299d8ad6.slice/crio-f31fd391d6155585aa7333ec2d143efe892b22b28a4240911ce1ec7c93027ffe WatchSource:0}: Error finding container f31fd391d6155585aa7333ec2d143efe892b22b28a4240911ce1ec7c93027ffe: Status 404 returned error can't find the container with id f31fd391d6155585aa7333ec2d143efe892b22b28a4240911ce1ec7c93027ffe Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.908002 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kmgld"] Dec 04 14:12:28 crc kubenswrapper[4848]: W1204 14:12:28.936633 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18c2d4b9_f05f_4eae_8cdc_a809508dd9e8.slice/crio-d750a4fc31520c3590c411f3835a62bfd04c3245720e29cf9cfcbb9421c8ae78 WatchSource:0}: Error finding container d750a4fc31520c3590c411f3835a62bfd04c3245720e29cf9cfcbb9421c8ae78: Status 404 returned error can't find the container with id d750a4fc31520c3590c411f3835a62bfd04c3245720e29cf9cfcbb9421c8ae78 Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.966500 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7b6b4654c9-hg8j8"] Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.968381 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.978342 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.978642 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 04 14:12:28 crc kubenswrapper[4848]: I1204 14:12:28.989922 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b6b4654c9-hg8j8"] Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.049247 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cc6cc5676-4j9jn"] Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.089574 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-combined-ca-bundle\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.089628 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-public-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.089648 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-config\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.089692 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-internal-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.089712 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pdff\" (UniqueName: \"kubernetes.io/projected/59773efa-83b9-4645-8187-fc0ba4c1d70c-kube-api-access-5pdff\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.089772 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-httpd-config\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.089808 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-ovndb-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.199238 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-internal-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.199288 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pdff\" (UniqueName: \"kubernetes.io/projected/59773efa-83b9-4645-8187-fc0ba4c1d70c-kube-api-access-5pdff\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.199363 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-httpd-config\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.199404 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-ovndb-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.199477 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-combined-ca-bundle\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.199511 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-config\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.199527 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-public-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.208008 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-httpd-config\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.210584 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-public-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.211079 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-internal-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.218319 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-config\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.218537 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-ovndb-tls-certs\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.220633 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59773efa-83b9-4645-8187-fc0ba4c1d70c-combined-ca-bundle\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.249713 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pdff\" (UniqueName: \"kubernetes.io/projected/59773efa-83b9-4645-8187-fc0ba4c1d70c-kube-api-access-5pdff\") pod \"neutron-7b6b4654c9-hg8j8\" (UID: \"59773efa-83b9-4645-8187-fc0ba4c1d70c\") " pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.309431 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.336746 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99e49830-7634-4257-af28-938f373af926","Type":"ContainerStarted","Data":"68899fbbe0a11f00377c19314d4e5f511d7f1555f358fbe9cf6d5159de5965f1"} Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.358171 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9r82" event={"ID":"deb8ffb1-2a45-4591-b749-ef7c9b2425bb","Type":"ContainerStarted","Data":"570461ea30f63038f02b2b2ddb9a7ccd81b193ba9bfe54003f38f4e9c768ceb1"} Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.358220 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9r82" event={"ID":"deb8ffb1-2a45-4591-b749-ef7c9b2425bb","Type":"ContainerStarted","Data":"8c1b51356b2c369e1baeb1ba3c4a45eb85fd206fa6e90d82a8c712bb34f25ace"} Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.376586 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f252353f-4af7-445f-9860-30d4299d8ad6","Type":"ContainerStarted","Data":"f31fd391d6155585aa7333ec2d143efe892b22b28a4240911ce1ec7c93027ffe"} Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.382549 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc6cc5676-4j9jn" event={"ID":"9ad8065e-d54e-469b-bb54-f3a0ab26edf6","Type":"ContainerStarted","Data":"a2215a2024a3cb6bc1c5957115d2c6565d7e32c1fdcd3ea79fb38b2bd7f46dbf"} Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.385200 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" event={"ID":"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8","Type":"ContainerStarted","Data":"d750a4fc31520c3590c411f3835a62bfd04c3245720e29cf9cfcbb9421c8ae78"} Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.404400 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-t9r82" podStartSLOduration=14.404386311 podStartE2EDuration="14.404386311s" podCreationTimestamp="2025-12-04 14:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:29.390643068 +0000 UTC m=+1453.333139596" watchObservedRunningTime="2025-12-04 14:12:29.404386311 +0000 UTC m=+1453.346882839" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.429564 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-l5jbz" event={"ID":"23eae77d-63ca-4004-969b-cfac163d3975","Type":"ContainerStarted","Data":"092b1f360bd7cd0348b2807080cbd29d7a7ba4fc860af7e4bb8a684d87b00a6e"} Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.470772 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-l5jbz" podStartSLOduration=3.767605566 podStartE2EDuration="40.470752761s" podCreationTimestamp="2025-12-04 14:11:49 +0000 UTC" firstStartedPulling="2025-12-04 14:11:51.520184551 +0000 UTC m=+1415.462681079" lastFinishedPulling="2025-12-04 14:12:28.223331746 +0000 UTC m=+1452.165828274" observedRunningTime="2025-12-04 14:12:29.463163416 +0000 UTC m=+1453.405659944" watchObservedRunningTime="2025-12-04 14:12:29.470752761 +0000 UTC m=+1453.413249289" Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.488185 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:12:29 crc kubenswrapper[4848]: I1204 14:12:29.642081 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-x9k96" podUID="c8675f3d-03ac-46e5-be9e-84991aa3e80a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: i/o timeout" Dec 04 14:12:30 crc kubenswrapper[4848]: I1204 14:12:30.082996 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b6b4654c9-hg8j8"] Dec 04 14:12:30 crc kubenswrapper[4848]: I1204 14:12:30.473568 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db70a1ef-eaa8-4da0-8abe-72aa5c822226","Type":"ContainerStarted","Data":"bc6faf4dc722e4aaedc09369cabab98084d15e87c650dbb55ad5712971365e06"} Dec 04 14:12:30 crc kubenswrapper[4848]: I1204 14:12:30.477024 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b6b4654c9-hg8j8" event={"ID":"59773efa-83b9-4645-8187-fc0ba4c1d70c","Type":"ContainerStarted","Data":"23783009bc5348c38148b0d1f95dc6361147398bd0bf78d4039eebf1c01d72df"} Dec 04 14:12:31 crc kubenswrapper[4848]: I1204 14:12:31.488916 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc6cc5676-4j9jn" event={"ID":"9ad8065e-d54e-469b-bb54-f3a0ab26edf6","Type":"ContainerStarted","Data":"2c5301185af0319c217c4f0b837fa82d49f72885cc8d9695ba56357b0224bda3"} Dec 04 14:12:32 crc kubenswrapper[4848]: I1204 14:12:32.500195 4848 generic.go:334] "Generic (PLEG): container finished" podID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerID="200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0" exitCode=0 Dec 04 14:12:32 crc kubenswrapper[4848]: I1204 14:12:32.500268 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" event={"ID":"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8","Type":"ContainerDied","Data":"200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0"} Dec 04 14:12:32 crc kubenswrapper[4848]: I1204 14:12:32.503772 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f252353f-4af7-445f-9860-30d4299d8ad6","Type":"ContainerStarted","Data":"3befd5657f14087a56be9aa7c8ccf6b25b52ce5f18f74cc76d3b81e1e72396ae"} Dec 04 14:12:32 crc kubenswrapper[4848]: I1204 14:12:32.505576 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db70a1ef-eaa8-4da0-8abe-72aa5c822226","Type":"ContainerStarted","Data":"c7727796ae2e569ffb4887e1d8acaaae87939cc44b10b82cc3a84810cd2191aa"} Dec 04 14:12:32 crc kubenswrapper[4848]: I1204 14:12:32.507205 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b6b4654c9-hg8j8" event={"ID":"59773efa-83b9-4645-8187-fc0ba4c1d70c","Type":"ContainerStarted","Data":"99db83afc7f686f1ec3c3e6211c93c2d448ba05c217535434f712a5c9dda967a"} Dec 04 14:12:33 crc kubenswrapper[4848]: I1204 14:12:33.518843 4848 generic.go:334] "Generic (PLEG): container finished" podID="deb8ffb1-2a45-4591-b749-ef7c9b2425bb" containerID="570461ea30f63038f02b2b2ddb9a7ccd81b193ba9bfe54003f38f4e9c768ceb1" exitCode=0 Dec 04 14:12:33 crc kubenswrapper[4848]: I1204 14:12:33.518919 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9r82" event={"ID":"deb8ffb1-2a45-4591-b749-ef7c9b2425bb","Type":"ContainerDied","Data":"570461ea30f63038f02b2b2ddb9a7ccd81b193ba9bfe54003f38f4e9c768ceb1"} Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.448053 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.583798 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-scripts\") pod \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.584080 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-credential-keys\") pod \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.584107 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-config-data\") pod \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.584190 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-combined-ca-bundle\") pod \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.584236 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzt65\" (UniqueName: \"kubernetes.io/projected/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-kube-api-access-jzt65\") pod \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.584293 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-fernet-keys\") pod \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\" (UID: \"deb8ffb1-2a45-4591-b749-ef7c9b2425bb\") " Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.591387 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-scripts" (OuterVolumeSpecName: "scripts") pod "deb8ffb1-2a45-4591-b749-ef7c9b2425bb" (UID: "deb8ffb1-2a45-4591-b749-ef7c9b2425bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.597393 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "deb8ffb1-2a45-4591-b749-ef7c9b2425bb" (UID: "deb8ffb1-2a45-4591-b749-ef7c9b2425bb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.600811 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t9r82" event={"ID":"deb8ffb1-2a45-4591-b749-ef7c9b2425bb","Type":"ContainerDied","Data":"8c1b51356b2c369e1baeb1ba3c4a45eb85fd206fa6e90d82a8c712bb34f25ace"} Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.600847 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c1b51356b2c369e1baeb1ba3c4a45eb85fd206fa6e90d82a8c712bb34f25ace" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.600855 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t9r82" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.603648 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-kube-api-access-jzt65" (OuterVolumeSpecName: "kube-api-access-jzt65") pod "deb8ffb1-2a45-4591-b749-ef7c9b2425bb" (UID: "deb8ffb1-2a45-4591-b749-ef7c9b2425bb"). InnerVolumeSpecName "kube-api-access-jzt65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.604586 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "deb8ffb1-2a45-4591-b749-ef7c9b2425bb" (UID: "deb8ffb1-2a45-4591-b749-ef7c9b2425bb"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.634516 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "deb8ffb1-2a45-4591-b749-ef7c9b2425bb" (UID: "deb8ffb1-2a45-4591-b749-ef7c9b2425bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.649654 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-config-data" (OuterVolumeSpecName: "config-data") pod "deb8ffb1-2a45-4591-b749-ef7c9b2425bb" (UID: "deb8ffb1-2a45-4591-b749-ef7c9b2425bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.688079 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.688113 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.688126 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzt65\" (UniqueName: \"kubernetes.io/projected/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-kube-api-access-jzt65\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.688135 4848 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.688145 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:36 crc kubenswrapper[4848]: I1204 14:12:36.688153 4848 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/deb8ffb1-2a45-4591-b749-ef7c9b2425bb-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.620233 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db70a1ef-eaa8-4da0-8abe-72aa5c822226","Type":"ContainerStarted","Data":"099fcfe1d45f56ac6d66db57942f93cf89b334794714da0f9c2e4368156413a1"} Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.627643 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b6b4654c9-hg8j8" event={"ID":"59773efa-83b9-4645-8187-fc0ba4c1d70c","Type":"ContainerStarted","Data":"438972e6eb22342f75888916427f98b5f1a3b3b9b54885f2b90fe1271c6b4441"} Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.627795 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.630780 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99e49830-7634-4257-af28-938f373af926","Type":"ContainerStarted","Data":"841d00b71ea982d6c3ddac488a7d77982f117bb6214829ecfacc33e54fa5011c"} Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.635182 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc6cc5676-4j9jn" event={"ID":"9ad8065e-d54e-469b-bb54-f3a0ab26edf6","Type":"ContainerStarted","Data":"5c95a5666dc776ffb5d5e6b3b266755d63c44b32ccbb3adf8b9502998c49c994"} Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.636282 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.638662 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-m9h4t" event={"ID":"7380f023-2712-4a2b-a193-ec401d07c9ad","Type":"ContainerStarted","Data":"f280ccc815c96ade43f1fc073378ab7a2f03d67964c55418b44f3d5c2300f082"} Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.657301 4848 generic.go:334] "Generic (PLEG): container finished" podID="23eae77d-63ca-4004-969b-cfac163d3975" containerID="092b1f360bd7cd0348b2807080cbd29d7a7ba4fc860af7e4bb8a684d87b00a6e" exitCode=0 Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.657370 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-l5jbz" event={"ID":"23eae77d-63ca-4004-969b-cfac163d3975","Type":"ContainerDied","Data":"092b1f360bd7cd0348b2807080cbd29d7a7ba4fc860af7e4bb8a684d87b00a6e"} Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.660185 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" event={"ID":"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8","Type":"ContainerStarted","Data":"15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22"} Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.661021 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.664438 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f252353f-4af7-445f-9860-30d4299d8ad6","Type":"ContainerStarted","Data":"5c35ca14f75f66593173d79a1d0a76f32adfb5dbc514ced5be9ab4b76b525913"} Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.682179 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.682157711 podStartE2EDuration="11.682157711s" podCreationTimestamp="2025-12-04 14:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:37.653493297 +0000 UTC m=+1461.595989835" watchObservedRunningTime="2025-12-04 14:12:37.682157711 +0000 UTC m=+1461.624654239" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.692047 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7b6b4654c9-hg8j8" podStartSLOduration=9.692024931 podStartE2EDuration="9.692024931s" podCreationTimestamp="2025-12-04 14:12:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:37.676459183 +0000 UTC m=+1461.618955711" watchObservedRunningTime="2025-12-04 14:12:37.692024931 +0000 UTC m=+1461.634521469" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.720810 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cc6cc5676-4j9jn" podStartSLOduration=11.720790269 podStartE2EDuration="11.720790269s" podCreationTimestamp="2025-12-04 14:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:37.697887454 +0000 UTC m=+1461.640383982" watchObservedRunningTime="2025-12-04 14:12:37.720790269 +0000 UTC m=+1461.663286797" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.765886 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5b74ff8c74-nm59b"] Dec 04 14:12:37 crc kubenswrapper[4848]: E1204 14:12:37.766634 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb8ffb1-2a45-4591-b749-ef7c9b2425bb" containerName="keystone-bootstrap" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.766660 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb8ffb1-2a45-4591-b749-ef7c9b2425bb" containerName="keystone-bootstrap" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.766883 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb8ffb1-2a45-4591-b749-ef7c9b2425bb" containerName="keystone-bootstrap" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.767776 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.772315 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.772471 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.772573 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.772683 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.772873 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4527c" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.781838 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-m9h4t" podStartSLOduration=4.186148228 podStartE2EDuration="48.781793129s" podCreationTimestamp="2025-12-04 14:11:49 +0000 UTC" firstStartedPulling="2025-12-04 14:11:51.921787692 +0000 UTC m=+1415.864284220" lastFinishedPulling="2025-12-04 14:12:36.517432593 +0000 UTC m=+1460.459929121" observedRunningTime="2025-12-04 14:12:37.724112769 +0000 UTC m=+1461.666609307" watchObservedRunningTime="2025-12-04 14:12:37.781793129 +0000 UTC m=+1461.724289657" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.791230 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.814292 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b74ff8c74-nm59b"] Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.856272 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=20.856248824 podStartE2EDuration="20.856248824s" podCreationTimestamp="2025-12-04 14:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:37.790206773 +0000 UTC m=+1461.732703301" watchObservedRunningTime="2025-12-04 14:12:37.856248824 +0000 UTC m=+1461.798745372" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.864207 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" podStartSLOduration=11.864190196 podStartE2EDuration="11.864190196s" podCreationTimestamp="2025-12-04 14:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:37.828760477 +0000 UTC m=+1461.771257005" watchObservedRunningTime="2025-12-04 14:12:37.864190196 +0000 UTC m=+1461.806686724" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.918506 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-combined-ca-bundle\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.918584 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-public-tls-certs\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.918616 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-credential-keys\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.918697 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-config-data\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.918742 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmzvp\" (UniqueName: \"kubernetes.io/projected/6a28b540-21e3-43eb-9b64-215d661d4721-kube-api-access-fmzvp\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.918794 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-scripts\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.918867 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-fernet-keys\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:37 crc kubenswrapper[4848]: I1204 14:12:37.918894 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-internal-tls-certs\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.021100 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-public-tls-certs\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.021438 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-credential-keys\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.021512 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-config-data\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.021569 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmzvp\" (UniqueName: \"kubernetes.io/projected/6a28b540-21e3-43eb-9b64-215d661d4721-kube-api-access-fmzvp\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.021782 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-scripts\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.021862 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-fernet-keys\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.021899 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-internal-tls-certs\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.022048 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-combined-ca-bundle\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.027927 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-fernet-keys\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.028205 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-combined-ca-bundle\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.028519 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-config-data\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.033540 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-internal-tls-certs\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.037330 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-scripts\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.038930 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-public-tls-certs\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.043353 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a28b540-21e3-43eb-9b64-215d661d4721-credential-keys\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.045238 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmzvp\" (UniqueName: \"kubernetes.io/projected/6a28b540-21e3-43eb-9b64-215d661d4721-kube-api-access-fmzvp\") pod \"keystone-5b74ff8c74-nm59b\" (UID: \"6a28b540-21e3-43eb-9b64-215d661d4721\") " pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.104066 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:38 crc kubenswrapper[4848]: I1204 14:12:38.698984 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b74ff8c74-nm59b"] Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.060419 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.159643 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-db-sync-config-data\") pod \"23eae77d-63ca-4004-969b-cfac163d3975\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.159767 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpl44\" (UniqueName: \"kubernetes.io/projected/23eae77d-63ca-4004-969b-cfac163d3975-kube-api-access-xpl44\") pod \"23eae77d-63ca-4004-969b-cfac163d3975\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.159816 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-combined-ca-bundle\") pod \"23eae77d-63ca-4004-969b-cfac163d3975\" (UID: \"23eae77d-63ca-4004-969b-cfac163d3975\") " Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.164308 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23eae77d-63ca-4004-969b-cfac163d3975-kube-api-access-xpl44" (OuterVolumeSpecName: "kube-api-access-xpl44") pod "23eae77d-63ca-4004-969b-cfac163d3975" (UID: "23eae77d-63ca-4004-969b-cfac163d3975"). InnerVolumeSpecName "kube-api-access-xpl44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.164605 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "23eae77d-63ca-4004-969b-cfac163d3975" (UID: "23eae77d-63ca-4004-969b-cfac163d3975"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.187174 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23eae77d-63ca-4004-969b-cfac163d3975" (UID: "23eae77d-63ca-4004-969b-cfac163d3975"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.262075 4848 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.262109 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpl44\" (UniqueName: \"kubernetes.io/projected/23eae77d-63ca-4004-969b-cfac163d3975-kube-api-access-xpl44\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.262121 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eae77d-63ca-4004-969b-cfac163d3975-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.684734 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-l5jbz" event={"ID":"23eae77d-63ca-4004-969b-cfac163d3975","Type":"ContainerDied","Data":"a81e9d511512ada4ab85d1225e9043ed8750085cee32fb0001d67566e0fd29ca"} Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.684771 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a81e9d511512ada4ab85d1225e9043ed8750085cee32fb0001d67566e0fd29ca" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.684747 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-l5jbz" Dec 04 14:12:39 crc kubenswrapper[4848]: I1204 14:12:39.686756 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b74ff8c74-nm59b" event={"ID":"6a28b540-21e3-43eb-9b64-215d661d4721","Type":"ContainerStarted","Data":"5b8905a906037353d35916ea592e16e38e8fbd71946a599c36ae88cc468868e2"} Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.039840 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7b699c955-dtjdb"] Dec 04 14:12:40 crc kubenswrapper[4848]: E1204 14:12:40.040354 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23eae77d-63ca-4004-969b-cfac163d3975" containerName="barbican-db-sync" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.040366 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="23eae77d-63ca-4004-969b-cfac163d3975" containerName="barbican-db-sync" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.040582 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="23eae77d-63ca-4004-969b-cfac163d3975" containerName="barbican-db-sync" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.041683 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.043861 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.043971 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-q6vwc" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.044188 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.059010 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b699c955-dtjdb"] Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.160220 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-58957fbb8b-n55cg"] Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.163773 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.174335 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.188253 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-config-data\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.188356 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgshk\" (UniqueName: \"kubernetes.io/projected/051c8366-c751-423e-a7c8-4d69b6a5897d-kube-api-access-rgshk\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.188394 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-combined-ca-bundle\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.188438 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-config-data-custom\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.188484 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/051c8366-c751-423e-a7c8-4d69b6a5897d-logs\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.201324 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-58957fbb8b-n55cg"] Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.224855 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kmgld"] Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.233744 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pk46n"] Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.235534 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.252402 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pk46n"] Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.302723 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-config-data-custom\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.302790 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c232050c-845f-4a98-b79d-1788830cfd67-logs\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.302811 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-combined-ca-bundle\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.302853 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-config-data\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.302891 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgshk\" (UniqueName: \"kubernetes.io/projected/051c8366-c751-423e-a7c8-4d69b6a5897d-kube-api-access-rgshk\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.302913 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-combined-ca-bundle\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.303026 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-config-data-custom\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.303069 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/051c8366-c751-423e-a7c8-4d69b6a5897d-logs\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.303109 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl2z7\" (UniqueName: \"kubernetes.io/projected/c232050c-845f-4a98-b79d-1788830cfd67-kube-api-access-vl2z7\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.303141 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-config-data\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.304366 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/051c8366-c751-423e-a7c8-4d69b6a5897d-logs\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.309753 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6fb4f55966-dj9bq"] Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.309967 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-config-data\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.310812 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-combined-ca-bundle\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.311641 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.314195 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.324664 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgshk\" (UniqueName: \"kubernetes.io/projected/051c8366-c751-423e-a7c8-4d69b6a5897d-kube-api-access-rgshk\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.326902 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6fb4f55966-dj9bq"] Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.339612 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/051c8366-c751-423e-a7c8-4d69b6a5897d-config-data-custom\") pod \"barbican-worker-7b699c955-dtjdb\" (UID: \"051c8366-c751-423e-a7c8-4d69b6a5897d\") " pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.390131 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b699c955-dtjdb" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.405570 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-svc\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406266 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9bjb\" (UniqueName: \"kubernetes.io/projected/9aa785c8-735e-42ff-a2dc-29481f49c4a3-kube-api-access-z9bjb\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406367 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406438 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa785c8-735e-42ff-a2dc-29481f49c4a3-logs\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406488 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406520 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl2z7\" (UniqueName: \"kubernetes.io/projected/c232050c-845f-4a98-b79d-1788830cfd67-kube-api-access-vl2z7\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406578 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-config-data\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406599 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data-custom\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406921 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdp2t\" (UniqueName: \"kubernetes.io/projected/bb3fe94e-7191-41a1-91a8-bd35ebd781de-kube-api-access-mdp2t\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.406962 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-combined-ca-bundle\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.407080 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-config-data-custom\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.407140 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.407199 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c232050c-845f-4a98-b79d-1788830cfd67-logs\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.407218 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-combined-ca-bundle\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.407236 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-config\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.407251 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.408383 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c232050c-845f-4a98-b79d-1788830cfd67-logs\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.411931 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-config-data-custom\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.413824 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-combined-ca-bundle\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.418832 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c232050c-845f-4a98-b79d-1788830cfd67-config-data\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.430073 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl2z7\" (UniqueName: \"kubernetes.io/projected/c232050c-845f-4a98-b79d-1788830cfd67-kube-api-access-vl2z7\") pod \"barbican-keystone-listener-58957fbb8b-n55cg\" (UID: \"c232050c-845f-4a98-b79d-1788830cfd67\") " pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.508799 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509091 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-config\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509111 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509171 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-svc\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509244 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9bjb\" (UniqueName: \"kubernetes.io/projected/9aa785c8-735e-42ff-a2dc-29481f49c4a3-kube-api-access-z9bjb\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509272 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509313 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa785c8-735e-42ff-a2dc-29481f49c4a3-logs\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509345 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509381 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data-custom\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509411 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdp2t\" (UniqueName: \"kubernetes.io/projected/bb3fe94e-7191-41a1-91a8-bd35ebd781de-kube-api-access-mdp2t\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509425 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-combined-ca-bundle\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.509992 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.510384 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.510517 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa785c8-735e-42ff-a2dc-29481f49c4a3-logs\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.510537 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-config\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.511047 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.511554 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-svc\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.515077 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.515865 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-combined-ca-bundle\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.517181 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data-custom\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.524946 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.529532 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9bjb\" (UniqueName: \"kubernetes.io/projected/9aa785c8-735e-42ff-a2dc-29481f49c4a3-kube-api-access-z9bjb\") pod \"barbican-api-6fb4f55966-dj9bq\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.529628 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdp2t\" (UniqueName: \"kubernetes.io/projected/bb3fe94e-7191-41a1-91a8-bd35ebd781de-kube-api-access-mdp2t\") pod \"dnsmasq-dns-85ff748b95-pk46n\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.581432 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.737624 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" podUID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerName="dnsmasq-dns" containerID="cri-o://15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22" gracePeriod=10 Dec 04 14:12:40 crc kubenswrapper[4848]: I1204 14:12:40.822349 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:40.999705 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b699c955-dtjdb"] Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:41.454699 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6fb4f55966-dj9bq"] Dec 04 14:12:41 crc kubenswrapper[4848]: W1204 14:12:41.459842 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aa785c8_735e_42ff_a2dc_29481f49c4a3.slice/crio-83e2a08bf38292daa3da61eb077ebe2530ef00dcc60a708c39668aedddfa6553 WatchSource:0}: Error finding container 83e2a08bf38292daa3da61eb077ebe2530ef00dcc60a708c39668aedddfa6553: Status 404 returned error can't find the container with id 83e2a08bf38292daa3da61eb077ebe2530ef00dcc60a708c39668aedddfa6553 Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:41.468825 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-58957fbb8b-n55cg"] Dec 04 14:12:41 crc kubenswrapper[4848]: W1204 14:12:41.601296 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb3fe94e_7191_41a1_91a8_bd35ebd781de.slice/crio-ee8c337807c8db36393f21a3e7f7f965524bfc8b311e3f0194e1262eac091d85 WatchSource:0}: Error finding container ee8c337807c8db36393f21a3e7f7f965524bfc8b311e3f0194e1262eac091d85: Status 404 returned error can't find the container with id ee8c337807c8db36393f21a3e7f7f965524bfc8b311e3f0194e1262eac091d85 Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:41.603087 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pk46n"] Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:41.750212 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b699c955-dtjdb" event={"ID":"051c8366-c751-423e-a7c8-4d69b6a5897d","Type":"ContainerStarted","Data":"37f2b24ae9049b565fa385d5a5f711dcd95df4eed349b711cc8aaa902c15e73c"} Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:41.752157 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fb4f55966-dj9bq" event={"ID":"9aa785c8-735e-42ff-a2dc-29481f49c4a3","Type":"ContainerStarted","Data":"83e2a08bf38292daa3da61eb077ebe2530ef00dcc60a708c39668aedddfa6553"} Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:41.753094 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" event={"ID":"bb3fe94e-7191-41a1-91a8-bd35ebd781de","Type":"ContainerStarted","Data":"ee8c337807c8db36393f21a3e7f7f965524bfc8b311e3f0194e1262eac091d85"} Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:41.754522 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" event={"ID":"c232050c-845f-4a98-b79d-1788830cfd67","Type":"ContainerStarted","Data":"d837980bc34381249967518875ac5af4a15450fb1c0450cfe42b52eed0d478aa"} Dec 04 14:12:41 crc kubenswrapper[4848]: I1204 14:12:41.950590 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" podUID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.189:5353: connect: connection refused" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.634227 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.795434 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-nb\") pod \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.801548 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95bpb\" (UniqueName: \"kubernetes.io/projected/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-kube-api-access-95bpb\") pod \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.801907 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-kube-api-access-95bpb" (OuterVolumeSpecName: "kube-api-access-95bpb") pod "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" (UID: "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8"). InnerVolumeSpecName "kube-api-access-95bpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.801939 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-sb\") pod \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.818905 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-config\") pod \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.819076 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-swift-storage-0\") pod \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.819179 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-svc\") pod \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\" (UID: \"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8\") " Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.838722 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95bpb\" (UniqueName: \"kubernetes.io/projected/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-kube-api-access-95bpb\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.909940 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" (UID: "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.926679 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" (UID: "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.935483 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fb4f55966-dj9bq" event={"ID":"9aa785c8-735e-42ff-a2dc-29481f49c4a3","Type":"ContainerStarted","Data":"5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f"} Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.935529 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fb4f55966-dj9bq" event={"ID":"9aa785c8-735e-42ff-a2dc-29481f49c4a3","Type":"ContainerStarted","Data":"dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1"} Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.935663 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.935701 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.945367 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.945412 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.960161 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-config" (OuterVolumeSpecName: "config") pod "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" (UID: "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.960871 4848 generic.go:334] "Generic (PLEG): container finished" podID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerID="15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22" exitCode=0 Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.961091 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" event={"ID":"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8","Type":"ContainerDied","Data":"15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22"} Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.961129 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" event={"ID":"18c2d4b9-f05f-4eae-8cdc-a809508dd9e8","Type":"ContainerDied","Data":"d750a4fc31520c3590c411f3835a62bfd04c3245720e29cf9cfcbb9421c8ae78"} Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.961150 4848 scope.go:117] "RemoveContainer" containerID="15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.961411 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-kmgld" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.981317 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b74ff8c74-nm59b" event={"ID":"6a28b540-21e3-43eb-9b64-215d661d4721","Type":"ContainerStarted","Data":"419a44b2852c2a4d70bd8c55eaeab40520ab6e28260cdcc1c97652037cecc846"} Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.985800 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.988549 4848 generic.go:334] "Generic (PLEG): container finished" podID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" containerID="4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8" exitCode=0 Dec 04 14:12:42 crc kubenswrapper[4848]: I1204 14:12:42.988605 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" event={"ID":"bb3fe94e-7191-41a1-91a8-bd35ebd781de","Type":"ContainerDied","Data":"4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8"} Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.004538 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" (UID: "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.017742 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6fb4f55966-dj9bq" podStartSLOduration=3.017721641 podStartE2EDuration="3.017721641s" podCreationTimestamp="2025-12-04 14:12:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:42.986208227 +0000 UTC m=+1466.928704775" watchObservedRunningTime="2025-12-04 14:12:43.017721641 +0000 UTC m=+1466.960218169" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.037863 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" (UID: "18c2d4b9-f05f-4eae-8cdc-a809508dd9e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.044689 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5b74ff8c74-nm59b" podStartSLOduration=6.044671375 podStartE2EDuration="6.044671375s" podCreationTimestamp="2025-12-04 14:12:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:43.031611188 +0000 UTC m=+1466.974107736" watchObservedRunningTime="2025-12-04 14:12:43.044671375 +0000 UTC m=+1466.987167903" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.047486 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.047515 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.047524 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.086759 4848 scope.go:117] "RemoveContainer" containerID="200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.104978 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-786894c88b-2c6rg"] Dec 04 14:12:43 crc kubenswrapper[4848]: E1204 14:12:43.116540 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerName="init" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.116588 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerName="init" Dec 04 14:12:43 crc kubenswrapper[4848]: E1204 14:12:43.116611 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerName="dnsmasq-dns" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.116616 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerName="dnsmasq-dns" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.116894 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" containerName="dnsmasq-dns" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.118007 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-786894c88b-2c6rg"] Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.118084 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.122364 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.122430 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.142190 4848 scope.go:117] "RemoveContainer" containerID="15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22" Dec 04 14:12:43 crc kubenswrapper[4848]: E1204 14:12:43.142983 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22\": container with ID starting with 15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22 not found: ID does not exist" containerID="15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.143007 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22"} err="failed to get container status \"15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22\": rpc error: code = NotFound desc = could not find container \"15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22\": container with ID starting with 15b286a842595c732b3341548566ecd0e545d8b3a91e79e729b91de3d9525c22 not found: ID does not exist" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.143043 4848 scope.go:117] "RemoveContainer" containerID="200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0" Dec 04 14:12:43 crc kubenswrapper[4848]: E1204 14:12:43.143233 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0\": container with ID starting with 200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0 not found: ID does not exist" containerID="200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.143249 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0"} err="failed to get container status \"200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0\": rpc error: code = NotFound desc = could not find container \"200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0\": container with ID starting with 200a452f15f55c01fde3b8138d5fd8c3af033157ca5bae8daade7cf2587e9bf0 not found: ID does not exist" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.251467 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-combined-ca-bundle\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.251558 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-config-data\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.251583 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-public-tls-certs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.251629 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-config-data-custom\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.251684 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e05a6e1-667b-44bd-a532-e91c2245e876-logs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.251761 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmwgc\" (UniqueName: \"kubernetes.io/projected/5e05a6e1-667b-44bd-a532-e91c2245e876-kube-api-access-fmwgc\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.251857 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-internal-tls-certs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.334053 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kmgld"] Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.357779 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e05a6e1-667b-44bd-a532-e91c2245e876-logs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.357825 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmwgc\" (UniqueName: \"kubernetes.io/projected/5e05a6e1-667b-44bd-a532-e91c2245e876-kube-api-access-fmwgc\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.357847 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-internal-tls-certs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.357956 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-combined-ca-bundle\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.358007 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-config-data\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.358032 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-public-tls-certs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.358074 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-config-data-custom\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.358684 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e05a6e1-667b-44bd-a532-e91c2245e876-logs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.364618 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-combined-ca-bundle\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.367292 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-internal-tls-certs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.372428 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-config-data-custom\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.377388 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-kmgld"] Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.379447 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-config-data\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.386490 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e05a6e1-667b-44bd-a532-e91c2245e876-public-tls-certs\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.388573 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmwgc\" (UniqueName: \"kubernetes.io/projected/5e05a6e1-667b-44bd-a532-e91c2245e876-kube-api-access-fmwgc\") pod \"barbican-api-786894c88b-2c6rg\" (UID: \"5e05a6e1-667b-44bd-a532-e91c2245e876\") " pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.458550 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:43 crc kubenswrapper[4848]: I1204 14:12:43.960531 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-786894c88b-2c6rg"] Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.002056 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ggd6n" event={"ID":"dba8b0d2-85d2-45c6-9612-0652e5dfdd33","Type":"ContainerStarted","Data":"0754a58093d928184ce8017dad3911af3c33cc4056ba464d5fbecdcf2fc4dbb2"} Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.008527 4848 generic.go:334] "Generic (PLEG): container finished" podID="7380f023-2712-4a2b-a193-ec401d07c9ad" containerID="f280ccc815c96ade43f1fc073378ab7a2f03d67964c55418b44f3d5c2300f082" exitCode=0 Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.008619 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-m9h4t" event={"ID":"7380f023-2712-4a2b-a193-ec401d07c9ad","Type":"ContainerDied","Data":"f280ccc815c96ade43f1fc073378ab7a2f03d67964c55418b44f3d5c2300f082"} Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.010483 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4dpj6" event={"ID":"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5","Type":"ContainerStarted","Data":"2f6081dfc3eaf61bd0b3a94ff130d966d45fb0be83fd461934d3e37ecd86f81c"} Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.013558 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" event={"ID":"bb3fe94e-7191-41a1-91a8-bd35ebd781de","Type":"ContainerStarted","Data":"d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56"} Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.013992 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.032201 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-ggd6n" podStartSLOduration=3.422398774 podStartE2EDuration="55.032154646s" podCreationTimestamp="2025-12-04 14:11:49 +0000 UTC" firstStartedPulling="2025-12-04 14:11:50.88149384 +0000 UTC m=+1414.823990368" lastFinishedPulling="2025-12-04 14:12:42.491249712 +0000 UTC m=+1466.433746240" observedRunningTime="2025-12-04 14:12:44.020876602 +0000 UTC m=+1467.963373130" watchObservedRunningTime="2025-12-04 14:12:44.032154646 +0000 UTC m=+1467.974651184" Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.053907 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" podStartSLOduration=4.053886993 podStartE2EDuration="4.053886993s" podCreationTimestamp="2025-12-04 14:12:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:44.039587706 +0000 UTC m=+1467.982084234" watchObservedRunningTime="2025-12-04 14:12:44.053886993 +0000 UTC m=+1467.996383521" Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.061827 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-4dpj6" podStartSLOduration=3.722067813 podStartE2EDuration="55.061804885s" podCreationTimestamp="2025-12-04 14:11:49 +0000 UTC" firstStartedPulling="2025-12-04 14:11:51.146383855 +0000 UTC m=+1415.088880383" lastFinishedPulling="2025-12-04 14:12:42.486120927 +0000 UTC m=+1466.428617455" observedRunningTime="2025-12-04 14:12:44.055194144 +0000 UTC m=+1467.997690672" watchObservedRunningTime="2025-12-04 14:12:44.061804885 +0000 UTC m=+1468.004301413" Dec 04 14:12:44 crc kubenswrapper[4848]: W1204 14:12:44.216005 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e05a6e1_667b_44bd_a532_e91c2245e876.slice/crio-fe440ba1f27f2333a72b1a74ccd0f8fa9be0067f0bd30fd9678502d20640f577 WatchSource:0}: Error finding container fe440ba1f27f2333a72b1a74ccd0f8fa9be0067f0bd30fd9678502d20640f577: Status 404 returned error can't find the container with id fe440ba1f27f2333a72b1a74ccd0f8fa9be0067f0bd30fd9678502d20640f577 Dec 04 14:12:44 crc kubenswrapper[4848]: I1204 14:12:44.407907 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18c2d4b9-f05f-4eae-8cdc-a809508dd9e8" path="/var/lib/kubelet/pods/18c2d4b9-f05f-4eae-8cdc-a809508dd9e8/volumes" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.039599 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-786894c88b-2c6rg" event={"ID":"5e05a6e1-667b-44bd-a532-e91c2245e876","Type":"ContainerStarted","Data":"b905588846316ee8888a3aed1cf8e1a93c3fc5d8c3fdef92fc5485691b9f207a"} Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.039962 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-786894c88b-2c6rg" event={"ID":"5e05a6e1-667b-44bd-a532-e91c2245e876","Type":"ContainerStarted","Data":"fe440ba1f27f2333a72b1a74ccd0f8fa9be0067f0bd30fd9678502d20640f577"} Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.041657 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" event={"ID":"c232050c-845f-4a98-b79d-1788830cfd67","Type":"ContainerStarted","Data":"9ec315fbef9d551aab7cf65ad9dd7857cd4817cd24a299771ecf74d72b49940f"} Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.047327 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b699c955-dtjdb" event={"ID":"051c8366-c751-423e-a7c8-4d69b6a5897d","Type":"ContainerStarted","Data":"2ddcbe7b5c87cd4a17cca0759fc656bc6f13a8d9d416a27f82b9ad61c7adbabe"} Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.504011 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-m9h4t" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.621855 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmkjn\" (UniqueName: \"kubernetes.io/projected/7380f023-2712-4a2b-a193-ec401d07c9ad-kube-api-access-vmkjn\") pod \"7380f023-2712-4a2b-a193-ec401d07c9ad\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.621968 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-combined-ca-bundle\") pod \"7380f023-2712-4a2b-a193-ec401d07c9ad\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.622034 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-scripts\") pod \"7380f023-2712-4a2b-a193-ec401d07c9ad\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.622097 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-config-data\") pod \"7380f023-2712-4a2b-a193-ec401d07c9ad\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.622176 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7380f023-2712-4a2b-a193-ec401d07c9ad-logs\") pod \"7380f023-2712-4a2b-a193-ec401d07c9ad\" (UID: \"7380f023-2712-4a2b-a193-ec401d07c9ad\") " Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.622968 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7380f023-2712-4a2b-a193-ec401d07c9ad-logs" (OuterVolumeSpecName: "logs") pod "7380f023-2712-4a2b-a193-ec401d07c9ad" (UID: "7380f023-2712-4a2b-a193-ec401d07c9ad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.628066 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7380f023-2712-4a2b-a193-ec401d07c9ad-kube-api-access-vmkjn" (OuterVolumeSpecName: "kube-api-access-vmkjn") pod "7380f023-2712-4a2b-a193-ec401d07c9ad" (UID: "7380f023-2712-4a2b-a193-ec401d07c9ad"). InnerVolumeSpecName "kube-api-access-vmkjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.636217 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-scripts" (OuterVolumeSpecName: "scripts") pod "7380f023-2712-4a2b-a193-ec401d07c9ad" (UID: "7380f023-2712-4a2b-a193-ec401d07c9ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.655762 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-config-data" (OuterVolumeSpecName: "config-data") pod "7380f023-2712-4a2b-a193-ec401d07c9ad" (UID: "7380f023-2712-4a2b-a193-ec401d07c9ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.661134 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7380f023-2712-4a2b-a193-ec401d07c9ad" (UID: "7380f023-2712-4a2b-a193-ec401d07c9ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.725161 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.725202 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7380f023-2712-4a2b-a193-ec401d07c9ad-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.725215 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmkjn\" (UniqueName: \"kubernetes.io/projected/7380f023-2712-4a2b-a193-ec401d07c9ad-kube-api-access-vmkjn\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.725231 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:45 crc kubenswrapper[4848]: I1204 14:12:45.725243 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7380f023-2712-4a2b-a193-ec401d07c9ad-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.084507 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b699c955-dtjdb" event={"ID":"051c8366-c751-423e-a7c8-4d69b6a5897d","Type":"ContainerStarted","Data":"308e4c43be4052d2bdcc219843e49e05c1a5e635d44a420a6bd549ce845fc98a"} Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.094244 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-786894c88b-2c6rg" event={"ID":"5e05a6e1-667b-44bd-a532-e91c2245e876","Type":"ContainerStarted","Data":"840851cd9f0238d02a27f11bb88e2f651c2105596fbcbd16aed9989b606d1918"} Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.095029 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.095143 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.101767 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-m9h4t" event={"ID":"7380f023-2712-4a2b-a193-ec401d07c9ad","Type":"ContainerDied","Data":"bdbe8f1defca2fbb457b9f3e5971657fef8967088fed05531c5409659b0aaf41"} Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.102006 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdbe8f1defca2fbb457b9f3e5971657fef8967088fed05531c5409659b0aaf41" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.102163 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-m9h4t" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.128154 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" event={"ID":"c232050c-845f-4a98-b79d-1788830cfd67","Type":"ContainerStarted","Data":"919d79fd2c597f2f815fe37c32ba936598897488089d865df32c2227567b42bc"} Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.139324 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7b699c955-dtjdb" podStartSLOduration=3.582260441 podStartE2EDuration="7.139298483s" podCreationTimestamp="2025-12-04 14:12:39 +0000 UTC" firstStartedPulling="2025-12-04 14:12:41.061024514 +0000 UTC m=+1465.003521042" lastFinishedPulling="2025-12-04 14:12:44.618062556 +0000 UTC m=+1468.560559084" observedRunningTime="2025-12-04 14:12:46.112378799 +0000 UTC m=+1470.054875327" watchObservedRunningTime="2025-12-04 14:12:46.139298483 +0000 UTC m=+1470.081795011" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.173298 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-786894c88b-2c6rg" podStartSLOduration=3.173273267 podStartE2EDuration="3.173273267s" podCreationTimestamp="2025-12-04 14:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:46.142276915 +0000 UTC m=+1470.084773443" watchObservedRunningTime="2025-12-04 14:12:46.173273267 +0000 UTC m=+1470.115769795" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.184621 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-58957fbb8b-n55cg" podStartSLOduration=3.057703731 podStartE2EDuration="6.184596561s" podCreationTimestamp="2025-12-04 14:12:40 +0000 UTC" firstStartedPulling="2025-12-04 14:12:41.466921588 +0000 UTC m=+1465.409418116" lastFinishedPulling="2025-12-04 14:12:44.593814418 +0000 UTC m=+1468.536310946" observedRunningTime="2025-12-04 14:12:46.180547253 +0000 UTC m=+1470.123043781" watchObservedRunningTime="2025-12-04 14:12:46.184596561 +0000 UTC m=+1470.127093099" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.220686 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8766cc68-d2824"] Dec 04 14:12:46 crc kubenswrapper[4848]: E1204 14:12:46.221565 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7380f023-2712-4a2b-a193-ec401d07c9ad" containerName="placement-db-sync" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.221591 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7380f023-2712-4a2b-a193-ec401d07c9ad" containerName="placement-db-sync" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.221991 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="7380f023-2712-4a2b-a193-ec401d07c9ad" containerName="placement-db-sync" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.223734 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.229776 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.230189 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-k5j2j" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.230343 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.230457 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.231080 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.249970 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8766cc68-d2824"] Dec 04 14:12:46 crc kubenswrapper[4848]: E1204 14:12:46.332232 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7380f023_2712_4a2b_a193_ec401d07c9ad.slice/crio-bdbe8f1defca2fbb457b9f3e5971657fef8967088fed05531c5409659b0aaf41\": RecentStats: unable to find data in memory cache]" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.348146 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-internal-tls-certs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.348193 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-scripts\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.348227 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c8a40cb-aac2-4a08-96f3-79c97ad82007-logs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.348256 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-public-tls-certs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.348377 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-combined-ca-bundle\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.348523 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p86tc\" (UniqueName: \"kubernetes.io/projected/7c8a40cb-aac2-4a08-96f3-79c97ad82007-kube-api-access-p86tc\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.348591 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-config-data\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.450706 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p86tc\" (UniqueName: \"kubernetes.io/projected/7c8a40cb-aac2-4a08-96f3-79c97ad82007-kube-api-access-p86tc\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.450823 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-config-data\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.450865 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-internal-tls-certs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.450889 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-scripts\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.450915 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c8a40cb-aac2-4a08-96f3-79c97ad82007-logs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.450940 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-public-tls-certs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.451034 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-combined-ca-bundle\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.451777 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c8a40cb-aac2-4a08-96f3-79c97ad82007-logs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.455935 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-scripts\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.458785 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-public-tls-certs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.458823 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-combined-ca-bundle\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.458922 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-internal-tls-certs\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.460323 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c8a40cb-aac2-4a08-96f3-79c97ad82007-config-data\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.474309 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p86tc\" (UniqueName: \"kubernetes.io/projected/7c8a40cb-aac2-4a08-96f3-79c97ad82007-kube-api-access-p86tc\") pod \"placement-8766cc68-d2824\" (UID: \"7c8a40cb-aac2-4a08-96f3-79c97ad82007\") " pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:46 crc kubenswrapper[4848]: I1204 14:12:46.583224 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.055350 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.055407 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.103410 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.124481 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.144251 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.144297 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.567596 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.567898 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.567911 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.567921 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.616191 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:47 crc kubenswrapper[4848]: I1204 14:12:47.639409 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:49 crc kubenswrapper[4848]: I1204 14:12:49.998558 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 14:12:49 crc kubenswrapper[4848]: I1204 14:12:49.998919 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.015693 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.199970 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.213781 4848 generic.go:334] "Generic (PLEG): container finished" podID="dba8b0d2-85d2-45c6-9612-0652e5dfdd33" containerID="0754a58093d928184ce8017dad3911af3c33cc4056ba464d5fbecdcf2fc4dbb2" exitCode=0 Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.214057 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ggd6n" event={"ID":"dba8b0d2-85d2-45c6-9612-0652e5dfdd33","Type":"ContainerDied","Data":"0754a58093d928184ce8017dad3911af3c33cc4056ba464d5fbecdcf2fc4dbb2"} Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.214532 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.243309 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.584238 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.686186 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-sx2nq"] Dec 04 14:12:50 crc kubenswrapper[4848]: I1204 14:12:50.686491 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" podUID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerName="dnsmasq-dns" containerID="cri-o://40c77f9f36a9e6dee29394145e40a1fd23c2d22573461a8ab118f9ad085233d3" gracePeriod=10 Dec 04 14:12:51 crc kubenswrapper[4848]: I1204 14:12:51.234689 4848 generic.go:334] "Generic (PLEG): container finished" podID="47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" containerID="2f6081dfc3eaf61bd0b3a94ff130d966d45fb0be83fd461934d3e37ecd86f81c" exitCode=0 Dec 04 14:12:51 crc kubenswrapper[4848]: I1204 14:12:51.234774 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4dpj6" event={"ID":"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5","Type":"ContainerDied","Data":"2f6081dfc3eaf61bd0b3a94ff130d966d45fb0be83fd461934d3e37ecd86f81c"} Dec 04 14:12:51 crc kubenswrapper[4848]: I1204 14:12:51.237558 4848 generic.go:334] "Generic (PLEG): container finished" podID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerID="40c77f9f36a9e6dee29394145e40a1fd23c2d22573461a8ab118f9ad085233d3" exitCode=0 Dec 04 14:12:51 crc kubenswrapper[4848]: I1204 14:12:51.237593 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" event={"ID":"4e1bd61d-621d-4c6a-84d6-5981897a8320","Type":"ContainerDied","Data":"40c77f9f36a9e6dee29394145e40a1fd23c2d22573461a8ab118f9ad085233d3"} Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.502440 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" podUID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.183:5353: connect: connection refused" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.773111 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.862620 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ggd6n" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.869435 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.936677 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-db-sync-config-data\") pod \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.936788 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzhzf\" (UniqueName: \"kubernetes.io/projected/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-kube-api-access-lzhzf\") pod \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.936824 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-etc-machine-id\") pod \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.936868 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-config-data\") pod \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.936883 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-combined-ca-bundle\") pod \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.937164 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p997\" (UniqueName: \"kubernetes.io/projected/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-kube-api-access-5p997\") pod \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.937204 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-combined-ca-bundle\") pod \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\" (UID: \"dba8b0d2-85d2-45c6-9612-0652e5dfdd33\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.937290 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" (UID: "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.937305 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-scripts\") pod \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.937394 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-config-data\") pod \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\" (UID: \"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5\") " Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.938156 4848 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.947111 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-scripts" (OuterVolumeSpecName: "scripts") pod "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" (UID: "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.962290 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" (UID: "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.962304 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-kube-api-access-5p997" (OuterVolumeSpecName: "kube-api-access-5p997") pod "dba8b0d2-85d2-45c6-9612-0652e5dfdd33" (UID: "dba8b0d2-85d2-45c6-9612-0652e5dfdd33"). InnerVolumeSpecName "kube-api-access-5p997". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.965121 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-kube-api-access-lzhzf" (OuterVolumeSpecName: "kube-api-access-lzhzf") pod "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" (UID: "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5"). InnerVolumeSpecName "kube-api-access-lzhzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:52 crc kubenswrapper[4848]: I1204 14:12:52.968777 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.018786 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dba8b0d2-85d2-45c6-9612-0652e5dfdd33" (UID: "dba8b0d2-85d2-45c6-9612-0652e5dfdd33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.038978 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.039010 4848 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.039021 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzhzf\" (UniqueName: \"kubernetes.io/projected/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-kube-api-access-lzhzf\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.039030 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p997\" (UniqueName: \"kubernetes.io/projected/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-kube-api-access-5p997\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.039039 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.040938 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" (UID: "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.041277 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-config-data" (OuterVolumeSpecName: "config-data") pod "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" (UID: "47da25b0-48bd-4dd1-8ad9-6db3f8565ad5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.116428 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-config-data" (OuterVolumeSpecName: "config-data") pod "dba8b0d2-85d2-45c6-9612-0652e5dfdd33" (UID: "dba8b0d2-85d2-45c6-9612-0652e5dfdd33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.143334 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.143362 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba8b0d2-85d2-45c6-9612-0652e5dfdd33-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.143371 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.283523 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ggd6n" event={"ID":"dba8b0d2-85d2-45c6-9612-0652e5dfdd33","Type":"ContainerDied","Data":"54ca3357c9f824eb53198308f66403f243da82dbacb0262d1cf73f921777696d"} Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.286897 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54ca3357c9f824eb53198308f66403f243da82dbacb0262d1cf73f921777696d" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.287102 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ggd6n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.308767 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4dpj6" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.309320 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4dpj6" event={"ID":"47da25b0-48bd-4dd1-8ad9-6db3f8565ad5","Type":"ContainerDied","Data":"1c1c180a6983b04cf02e03a08cf1b43c4584ffacd439497a5667613e5c029dc7"} Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.309375 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c1c180a6983b04cf02e03a08cf1b43c4584ffacd439497a5667613e5c029dc7" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.543985 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:12:53 crc kubenswrapper[4848]: E1204 14:12:53.544455 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" containerName="cinder-db-sync" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.544466 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" containerName="cinder-db-sync" Dec 04 14:12:53 crc kubenswrapper[4848]: E1204 14:12:53.544489 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dba8b0d2-85d2-45c6-9612-0652e5dfdd33" containerName="heat-db-sync" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.544497 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="dba8b0d2-85d2-45c6-9612-0652e5dfdd33" containerName="heat-db-sync" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.544706 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="dba8b0d2-85d2-45c6-9612-0652e5dfdd33" containerName="heat-db-sync" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.544723 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" containerName="cinder-db-sync" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.547642 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.552361 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.552436 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.552589 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-qdffl" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.552627 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.569921 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.599047 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kbh4n"] Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.601053 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.625224 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kbh4n"] Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658268 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658515 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658586 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658644 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-config\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658707 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-scripts\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658752 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658807 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658872 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658907 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e19d950-51e2-4e37-9734-30330e4af07d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.658996 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwx8r\" (UniqueName: \"kubernetes.io/projected/88b349db-6f99-4351-8a39-7dc1ba31496d-kube-api-access-hwx8r\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.659063 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f72wc\" (UniqueName: \"kubernetes.io/projected/9e19d950-51e2-4e37-9734-30330e4af07d-kube-api-access-f72wc\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.659112 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761015 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761122 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761153 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761178 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-config\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761239 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-scripts\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761272 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761315 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761347 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761375 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e19d950-51e2-4e37-9734-30330e4af07d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761422 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwx8r\" (UniqueName: \"kubernetes.io/projected/88b349db-6f99-4351-8a39-7dc1ba31496d-kube-api-access-hwx8r\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761458 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f72wc\" (UniqueName: \"kubernetes.io/projected/9e19d950-51e2-4e37-9734-30330e4af07d-kube-api-access-f72wc\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.761485 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.762252 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-config\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.762847 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.763406 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.764823 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e19d950-51e2-4e37-9734-30330e4af07d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.765609 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.766403 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.769834 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.772097 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.772532 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-scripts\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.779864 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.788708 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwx8r\" (UniqueName: \"kubernetes.io/projected/88b349db-6f99-4351-8a39-7dc1ba31496d-kube-api-access-hwx8r\") pod \"dnsmasq-dns-5c9776ccc5-kbh4n\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.798298 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f72wc\" (UniqueName: \"kubernetes.io/projected/9e19d950-51e2-4e37-9734-30330e4af07d-kube-api-access-f72wc\") pod \"cinder-scheduler-0\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.810524 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.815983 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.824614 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.833629 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.888338 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.956565 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.965023 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.965447 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95e4160a-61ff-4925-b263-b20b08e18663-logs\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.965499 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data-custom\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.965547 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgdc5\" (UniqueName: \"kubernetes.io/projected/95e4160a-61ff-4925-b263-b20b08e18663-kube-api-access-rgdc5\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.965593 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95e4160a-61ff-4925-b263-b20b08e18663-etc-machine-id\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.965613 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:53 crc kubenswrapper[4848]: I1204 14:12:53.965635 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-scripts\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.067917 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95e4160a-61ff-4925-b263-b20b08e18663-etc-machine-id\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.067974 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.068006 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-scripts\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.068079 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.068126 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95e4160a-61ff-4925-b263-b20b08e18663-logs\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.068169 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data-custom\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.068213 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgdc5\" (UniqueName: \"kubernetes.io/projected/95e4160a-61ff-4925-b263-b20b08e18663-kube-api-access-rgdc5\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.068585 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95e4160a-61ff-4925-b263-b20b08e18663-etc-machine-id\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.069834 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95e4160a-61ff-4925-b263-b20b08e18663-logs\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.072384 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.084824 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data-custom\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.086916 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgdc5\" (UniqueName: \"kubernetes.io/projected/95e4160a-61ff-4925-b263-b20b08e18663-kube-api-access-rgdc5\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.087434 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.099390 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-scripts\") pod \"cinder-api-0\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.214517 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.618514 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.784344 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-svc\") pod \"4e1bd61d-621d-4c6a-84d6-5981897a8320\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.784440 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-swift-storage-0\") pod \"4e1bd61d-621d-4c6a-84d6-5981897a8320\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.784462 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-nb\") pod \"4e1bd61d-621d-4c6a-84d6-5981897a8320\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.784585 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-sb\") pod \"4e1bd61d-621d-4c6a-84d6-5981897a8320\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.784612 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv6w5\" (UniqueName: \"kubernetes.io/projected/4e1bd61d-621d-4c6a-84d6-5981897a8320-kube-api-access-kv6w5\") pod \"4e1bd61d-621d-4c6a-84d6-5981897a8320\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.784634 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-config\") pod \"4e1bd61d-621d-4c6a-84d6-5981897a8320\" (UID: \"4e1bd61d-621d-4c6a-84d6-5981897a8320\") " Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.846307 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e1bd61d-621d-4c6a-84d6-5981897a8320-kube-api-access-kv6w5" (OuterVolumeSpecName: "kube-api-access-kv6w5") pod "4e1bd61d-621d-4c6a-84d6-5981897a8320" (UID: "4e1bd61d-621d-4c6a-84d6-5981897a8320"). InnerVolumeSpecName "kube-api-access-kv6w5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.892464 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv6w5\" (UniqueName: \"kubernetes.io/projected/4e1bd61d-621d-4c6a-84d6-5981897a8320-kube-api-access-kv6w5\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.914836 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4e1bd61d-621d-4c6a-84d6-5981897a8320" (UID: "4e1bd61d-621d-4c6a-84d6-5981897a8320"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.916891 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4e1bd61d-621d-4c6a-84d6-5981897a8320" (UID: "4e1bd61d-621d-4c6a-84d6-5981897a8320"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.929568 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4e1bd61d-621d-4c6a-84d6-5981897a8320" (UID: "4e1bd61d-621d-4c6a-84d6-5981897a8320"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.952471 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4e1bd61d-621d-4c6a-84d6-5981897a8320" (UID: "4e1bd61d-621d-4c6a-84d6-5981897a8320"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.969367 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-config" (OuterVolumeSpecName: "config") pod "4e1bd61d-621d-4c6a-84d6-5981897a8320" (UID: "4e1bd61d-621d-4c6a-84d6-5981897a8320"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.994853 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.994894 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.994909 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.994922 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:54 crc kubenswrapper[4848]: I1204 14:12:54.994936 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4e1bd61d-621d-4c6a-84d6-5981897a8320-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:55 crc kubenswrapper[4848]: E1204 14:12:55.260068 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="99e49830-7634-4257-af28-938f373af926" Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.330162 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" event={"ID":"4e1bd61d-621d-4c6a-84d6-5981897a8320","Type":"ContainerDied","Data":"9bf205f797a4aefb95d3d570284c9cc88faaafe11ffb93d2287667740b114358"} Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.330217 4848 scope.go:117] "RemoveContainer" containerID="40c77f9f36a9e6dee29394145e40a1fd23c2d22573461a8ab118f9ad085233d3" Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.330258 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-sx2nq" Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.336539 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99e49830-7634-4257-af28-938f373af926","Type":"ContainerStarted","Data":"615184e23c8aeb0ec7d6d6030f4b2dcd7cf611c05c94c19a1cf1b3569d831af2"} Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.336738 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99e49830-7634-4257-af28-938f373af926" containerName="ceilometer-notification-agent" containerID="cri-o://68899fbbe0a11f00377c19314d4e5f511d7f1555f358fbe9cf6d5159de5965f1" gracePeriod=30 Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.336801 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.336833 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99e49830-7634-4257-af28-938f373af926" containerName="proxy-httpd" containerID="cri-o://615184e23c8aeb0ec7d6d6030f4b2dcd7cf611c05c94c19a1cf1b3569d831af2" gracePeriod=30 Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.336905 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99e49830-7634-4257-af28-938f373af926" containerName="sg-core" containerID="cri-o://841d00b71ea982d6c3ddac488a7d77982f117bb6214829ecfacc33e54fa5011c" gracePeriod=30 Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.363625 4848 scope.go:117] "RemoveContainer" containerID="e0e35ed00f7715ec74f7a9d7407871e7d9c4652d57aa8a2ad307a9676d943187" Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.389426 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-sx2nq"] Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.399468 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-sx2nq"] Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.477397 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8766cc68-d2824"] Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.499653 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:12:55 crc kubenswrapper[4848]: W1204 14:12:55.860310 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e19d950_51e2_4e37_9734_30330e4af07d.slice/crio-d542ab438d63d7c7a94d7a7aa80a1b2c38a25af0a31a271aca659e64ba71f6e1 WatchSource:0}: Error finding container d542ab438d63d7c7a94d7a7aa80a1b2c38a25af0a31a271aca659e64ba71f6e1: Status 404 returned error can't find the container with id d542ab438d63d7c7a94d7a7aa80a1b2c38a25af0a31a271aca659e64ba71f6e1 Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.870721 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:12:55 crc kubenswrapper[4848]: I1204 14:12:55.881588 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kbh4n"] Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.291460 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.353404 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-786894c88b-2c6rg" Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.364357 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e19d950-51e2-4e37-9734-30330e4af07d","Type":"ContainerStarted","Data":"d542ab438d63d7c7a94d7a7aa80a1b2c38a25af0a31a271aca659e64ba71f6e1"} Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.371410 4848 generic.go:334] "Generic (PLEG): container finished" podID="99e49830-7634-4257-af28-938f373af926" containerID="841d00b71ea982d6c3ddac488a7d77982f117bb6214829ecfacc33e54fa5011c" exitCode=2 Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.371471 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99e49830-7634-4257-af28-938f373af926","Type":"ContainerDied","Data":"841d00b71ea982d6c3ddac488a7d77982f117bb6214829ecfacc33e54fa5011c"} Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.379880 4848 generic.go:334] "Generic (PLEG): container finished" podID="88b349db-6f99-4351-8a39-7dc1ba31496d" containerID="6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f" exitCode=0 Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.380025 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" event={"ID":"88b349db-6f99-4351-8a39-7dc1ba31496d","Type":"ContainerDied","Data":"6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f"} Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.380096 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" event={"ID":"88b349db-6f99-4351-8a39-7dc1ba31496d","Type":"ContainerStarted","Data":"3b5901029957843d3bd29a36cf59301d0d1cf99518d52242cf4731d06e30aa5f"} Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.505542 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e1bd61d-621d-4c6a-84d6-5981897a8320" path="/var/lib/kubelet/pods/4e1bd61d-621d-4c6a-84d6-5981897a8320/volumes" Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.506806 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"95e4160a-61ff-4925-b263-b20b08e18663","Type":"ContainerStarted","Data":"e69da11cdb84b1cb3eeb946e0928ccb1bd53530999ef411b1c8a4f044c3fea83"} Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.506840 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6fb4f55966-dj9bq"] Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.506863 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8766cc68-d2824" event={"ID":"7c8a40cb-aac2-4a08-96f3-79c97ad82007","Type":"ContainerStarted","Data":"e90bdc4309fce5b64d9f9bab999103617fa34b7899656e72d3cb2420a704202d"} Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.506883 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8766cc68-d2824" event={"ID":"7c8a40cb-aac2-4a08-96f3-79c97ad82007","Type":"ContainerStarted","Data":"73c561ccb0cc3cb92c65a9c887b17c80a0f567d95ad82caad84fd27a365bcbc5"} Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.507450 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6fb4f55966-dj9bq" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api-log" containerID="cri-o://dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1" gracePeriod=30 Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.512067 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6fb4f55966-dj9bq" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api" containerID="cri-o://5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f" gracePeriod=30 Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.521646 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6fb4f55966-dj9bq" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.196:9311/healthcheck\": EOF" Dec 04 14:12:56 crc kubenswrapper[4848]: E1204 14:12:56.811000 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aa785c8_735e_42ff_a2dc_29481f49c4a3.slice/crio-conmon-dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aa785c8_735e_42ff_a2dc_29481f49c4a3.slice/crio-dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:12:56 crc kubenswrapper[4848]: I1204 14:12:56.918775 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.251453 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.549263 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" event={"ID":"88b349db-6f99-4351-8a39-7dc1ba31496d","Type":"ContainerStarted","Data":"498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e"} Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.550902 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.571805 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" podStartSLOduration=4.571789338 podStartE2EDuration="4.571789338s" podCreationTimestamp="2025-12-04 14:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:57.566975861 +0000 UTC m=+1481.509472379" watchObservedRunningTime="2025-12-04 14:12:57.571789338 +0000 UTC m=+1481.514285866" Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.572790 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"95e4160a-61ff-4925-b263-b20b08e18663","Type":"ContainerStarted","Data":"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4"} Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.578789 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8766cc68-d2824" event={"ID":"7c8a40cb-aac2-4a08-96f3-79c97ad82007","Type":"ContainerStarted","Data":"d7aefd2bf629a531ef7fe88124a5876f0941827550291ce007b271b54c2ccd29"} Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.578847 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.578867 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8766cc68-d2824" Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.593758 4848 generic.go:334] "Generic (PLEG): container finished" podID="99e49830-7634-4257-af28-938f373af926" containerID="68899fbbe0a11f00377c19314d4e5f511d7f1555f358fbe9cf6d5159de5965f1" exitCode=0 Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.593841 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99e49830-7634-4257-af28-938f373af926","Type":"ContainerDied","Data":"68899fbbe0a11f00377c19314d4e5f511d7f1555f358fbe9cf6d5159de5965f1"} Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.605511 4848 generic.go:334] "Generic (PLEG): container finished" podID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerID="dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1" exitCode=143 Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.606018 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fb4f55966-dj9bq" event={"ID":"9aa785c8-735e-42ff-a2dc-29481f49c4a3","Type":"ContainerDied","Data":"dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1"} Dec 04 14:12:57 crc kubenswrapper[4848]: I1204 14:12:57.647277 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8766cc68-d2824" podStartSLOduration=11.647258269 podStartE2EDuration="11.647258269s" podCreationTimestamp="2025-12-04 14:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:57.634906529 +0000 UTC m=+1481.577403057" watchObservedRunningTime="2025-12-04 14:12:57.647258269 +0000 UTC m=+1481.589754797" Dec 04 14:12:58 crc kubenswrapper[4848]: I1204 14:12:58.624127 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"95e4160a-61ff-4925-b263-b20b08e18663","Type":"ContainerStarted","Data":"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2"} Dec 04 14:12:58 crc kubenswrapper[4848]: I1204 14:12:58.624314 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="95e4160a-61ff-4925-b263-b20b08e18663" containerName="cinder-api-log" containerID="cri-o://9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4" gracePeriod=30 Dec 04 14:12:58 crc kubenswrapper[4848]: I1204 14:12:58.624758 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 14:12:58 crc kubenswrapper[4848]: I1204 14:12:58.624807 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="95e4160a-61ff-4925-b263-b20b08e18663" containerName="cinder-api" containerID="cri-o://76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2" gracePeriod=30 Dec 04 14:12:58 crc kubenswrapper[4848]: I1204 14:12:58.628691 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e19d950-51e2-4e37-9734-30330e4af07d","Type":"ContainerStarted","Data":"e7ef746edbcd9791c05efb16f21feffebe8bd9e376a7c1fc8bac7d63f8a306fb"} Dec 04 14:12:58 crc kubenswrapper[4848]: I1204 14:12:58.656281 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.6562595810000005 podStartE2EDuration="5.656259581s" podCreationTimestamp="2025-12-04 14:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:12:58.648882762 +0000 UTC m=+1482.591379310" watchObservedRunningTime="2025-12-04 14:12:58.656259581 +0000 UTC m=+1482.598756099" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.329410 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7b6b4654c9-hg8j8" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.361075 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.450380 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cc6cc5676-4j9jn"] Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.450932 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cc6cc5676-4j9jn" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerName="neutron-api" containerID="cri-o://2c5301185af0319c217c4f0b837fa82d49f72885cc8d9695ba56357b0224bda3" gracePeriod=30 Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.451474 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cc6cc5676-4j9jn" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerName="neutron-httpd" containerID="cri-o://5c95a5666dc776ffb5d5e6b3b266755d63c44b32ccbb3adf8b9502998c49c994" gracePeriod=30 Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.511103 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95e4160a-61ff-4925-b263-b20b08e18663-logs\") pod \"95e4160a-61ff-4925-b263-b20b08e18663\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.511189 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95e4160a-61ff-4925-b263-b20b08e18663-etc-machine-id\") pod \"95e4160a-61ff-4925-b263-b20b08e18663\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.511214 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data-custom\") pod \"95e4160a-61ff-4925-b263-b20b08e18663\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.511317 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-combined-ca-bundle\") pod \"95e4160a-61ff-4925-b263-b20b08e18663\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.511376 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-scripts\") pod \"95e4160a-61ff-4925-b263-b20b08e18663\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.511394 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data\") pod \"95e4160a-61ff-4925-b263-b20b08e18663\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.511412 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgdc5\" (UniqueName: \"kubernetes.io/projected/95e4160a-61ff-4925-b263-b20b08e18663-kube-api-access-rgdc5\") pod \"95e4160a-61ff-4925-b263-b20b08e18663\" (UID: \"95e4160a-61ff-4925-b263-b20b08e18663\") " Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.513310 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95e4160a-61ff-4925-b263-b20b08e18663-logs" (OuterVolumeSpecName: "logs") pod "95e4160a-61ff-4925-b263-b20b08e18663" (UID: "95e4160a-61ff-4925-b263-b20b08e18663"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.513535 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95e4160a-61ff-4925-b263-b20b08e18663-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "95e4160a-61ff-4925-b263-b20b08e18663" (UID: "95e4160a-61ff-4925-b263-b20b08e18663"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.520073 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "95e4160a-61ff-4925-b263-b20b08e18663" (UID: "95e4160a-61ff-4925-b263-b20b08e18663"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.531547 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-scripts" (OuterVolumeSpecName: "scripts") pod "95e4160a-61ff-4925-b263-b20b08e18663" (UID: "95e4160a-61ff-4925-b263-b20b08e18663"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.536620 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e4160a-61ff-4925-b263-b20b08e18663-kube-api-access-rgdc5" (OuterVolumeSpecName: "kube-api-access-rgdc5") pod "95e4160a-61ff-4925-b263-b20b08e18663" (UID: "95e4160a-61ff-4925-b263-b20b08e18663"). InnerVolumeSpecName "kube-api-access-rgdc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.570882 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95e4160a-61ff-4925-b263-b20b08e18663" (UID: "95e4160a-61ff-4925-b263-b20b08e18663"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.583625 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data" (OuterVolumeSpecName: "config-data") pod "95e4160a-61ff-4925-b263-b20b08e18663" (UID: "95e4160a-61ff-4925-b263-b20b08e18663"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.613756 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95e4160a-61ff-4925-b263-b20b08e18663-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.613809 4848 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/95e4160a-61ff-4925-b263-b20b08e18663-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.613822 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.613833 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.613844 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.613854 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95e4160a-61ff-4925-b263-b20b08e18663-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.613867 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgdc5\" (UniqueName: \"kubernetes.io/projected/95e4160a-61ff-4925-b263-b20b08e18663-kube-api-access-rgdc5\") on node \"crc\" DevicePath \"\"" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.648274 4848 generic.go:334] "Generic (PLEG): container finished" podID="95e4160a-61ff-4925-b263-b20b08e18663" containerID="76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2" exitCode=0 Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.648312 4848 generic.go:334] "Generic (PLEG): container finished" podID="95e4160a-61ff-4925-b263-b20b08e18663" containerID="9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4" exitCode=143 Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.648358 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"95e4160a-61ff-4925-b263-b20b08e18663","Type":"ContainerDied","Data":"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2"} Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.648390 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"95e4160a-61ff-4925-b263-b20b08e18663","Type":"ContainerDied","Data":"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4"} Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.648402 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"95e4160a-61ff-4925-b263-b20b08e18663","Type":"ContainerDied","Data":"e69da11cdb84b1cb3eeb946e0928ccb1bd53530999ef411b1c8a4f044c3fea83"} Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.648419 4848 scope.go:117] "RemoveContainer" containerID="76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.648575 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.662931 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e19d950-51e2-4e37-9734-30330e4af07d","Type":"ContainerStarted","Data":"674a27ff95dcea4514c7e41e2c48a5b561c747cdd6f5d2e6fad57181047d1deb"} Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.688504 4848 scope.go:117] "RemoveContainer" containerID="9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.690208 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.798328596 podStartE2EDuration="6.690192218s" podCreationTimestamp="2025-12-04 14:12:53 +0000 UTC" firstStartedPulling="2025-12-04 14:12:55.918592991 +0000 UTC m=+1479.861089519" lastFinishedPulling="2025-12-04 14:12:56.810456613 +0000 UTC m=+1480.752953141" observedRunningTime="2025-12-04 14:12:59.681388895 +0000 UTC m=+1483.623885453" watchObservedRunningTime="2025-12-04 14:12:59.690192218 +0000 UTC m=+1483.632688746" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.719818 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.725250 4848 scope.go:117] "RemoveContainer" containerID="76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2" Dec 04 14:12:59 crc kubenswrapper[4848]: E1204 14:12:59.726207 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2\": container with ID starting with 76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2 not found: ID does not exist" containerID="76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.726241 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2"} err="failed to get container status \"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2\": rpc error: code = NotFound desc = could not find container \"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2\": container with ID starting with 76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2 not found: ID does not exist" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.726263 4848 scope.go:117] "RemoveContainer" containerID="9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4" Dec 04 14:12:59 crc kubenswrapper[4848]: E1204 14:12:59.727048 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4\": container with ID starting with 9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4 not found: ID does not exist" containerID="9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.727072 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4"} err="failed to get container status \"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4\": rpc error: code = NotFound desc = could not find container \"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4\": container with ID starting with 9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4 not found: ID does not exist" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.727091 4848 scope.go:117] "RemoveContainer" containerID="76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.727266 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2"} err="failed to get container status \"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2\": rpc error: code = NotFound desc = could not find container \"76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2\": container with ID starting with 76216d88ffdab85cfb9f5c067650341cc5a2d982b7d6e69b496e0a2adaffccc2 not found: ID does not exist" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.727295 4848 scope.go:117] "RemoveContainer" containerID="9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.729413 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4"} err="failed to get container status \"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4\": rpc error: code = NotFound desc = could not find container \"9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4\": container with ID starting with 9be85d5343bc978e275537d56ac2fb3a7501b08eee97aceba41e9378815a6bd4 not found: ID does not exist" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.739096 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.754100 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:12:59 crc kubenswrapper[4848]: E1204 14:12:59.754676 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerName="dnsmasq-dns" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.754698 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerName="dnsmasq-dns" Dec 04 14:12:59 crc kubenswrapper[4848]: E1204 14:12:59.754734 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e4160a-61ff-4925-b263-b20b08e18663" containerName="cinder-api-log" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.754743 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e4160a-61ff-4925-b263-b20b08e18663" containerName="cinder-api-log" Dec 04 14:12:59 crc kubenswrapper[4848]: E1204 14:12:59.754771 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerName="init" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.754780 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerName="init" Dec 04 14:12:59 crc kubenswrapper[4848]: E1204 14:12:59.754798 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e4160a-61ff-4925-b263-b20b08e18663" containerName="cinder-api" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.754804 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e4160a-61ff-4925-b263-b20b08e18663" containerName="cinder-api" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.755013 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e4160a-61ff-4925-b263-b20b08e18663" containerName="cinder-api-log" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.755062 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e4160a-61ff-4925-b263-b20b08e18663" containerName="cinder-api" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.755076 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e1bd61d-621d-4c6a-84d6-5981897a8320" containerName="dnsmasq-dns" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.756314 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.762551 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.762781 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.762801 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.779364 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.923621 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjskv\" (UniqueName: \"kubernetes.io/projected/e0893eec-f2cb-41b8-a301-005dbf326afe-kube-api-access-jjskv\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.923738 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-scripts\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.923767 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-config-data\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.923881 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0893eec-f2cb-41b8-a301-005dbf326afe-logs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.923908 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.924012 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0893eec-f2cb-41b8-a301-005dbf326afe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.924043 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.924087 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:12:59 crc kubenswrapper[4848]: I1204 14:12:59.924195 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.026827 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0893eec-f2cb-41b8-a301-005dbf326afe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.026879 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.026910 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.026986 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0893eec-f2cb-41b8-a301-005dbf326afe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.027008 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.027094 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjskv\" (UniqueName: \"kubernetes.io/projected/e0893eec-f2cb-41b8-a301-005dbf326afe-kube-api-access-jjskv\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.027157 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-scripts\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.027172 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-config-data\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.027285 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0893eec-f2cb-41b8-a301-005dbf326afe-logs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.027307 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.028227 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0893eec-f2cb-41b8-a301-005dbf326afe-logs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.031857 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.032531 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.036465 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-scripts\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.036749 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.036847 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.037445 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0893eec-f2cb-41b8-a301-005dbf326afe-config-data\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.055079 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjskv\" (UniqueName: \"kubernetes.io/projected/e0893eec-f2cb-41b8-a301-005dbf326afe-kube-api-access-jjskv\") pod \"cinder-api-0\" (UID: \"e0893eec-f2cb-41b8-a301-005dbf326afe\") " pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.081537 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.410650 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e4160a-61ff-4925-b263-b20b08e18663" path="/var/lib/kubelet/pods/95e4160a-61ff-4925-b263-b20b08e18663/volumes" Dec 04 14:13:00 crc kubenswrapper[4848]: W1204 14:13:00.639197 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0893eec_f2cb_41b8_a301_005dbf326afe.slice/crio-e5b9ab44f647ad8a94435fe3cee064a409ae2a2758ffb282dbb0c57c39acdc9e WatchSource:0}: Error finding container e5b9ab44f647ad8a94435fe3cee064a409ae2a2758ffb282dbb0c57c39acdc9e: Status 404 returned error can't find the container with id e5b9ab44f647ad8a94435fe3cee064a409ae2a2758ffb282dbb0c57c39acdc9e Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.644082 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.672606 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0893eec-f2cb-41b8-a301-005dbf326afe","Type":"ContainerStarted","Data":"e5b9ab44f647ad8a94435fe3cee064a409ae2a2758ffb282dbb0c57c39acdc9e"} Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.675061 4848 generic.go:334] "Generic (PLEG): container finished" podID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerID="5c95a5666dc776ffb5d5e6b3b266755d63c44b32ccbb3adf8b9502998c49c994" exitCode=0 Dec 04 14:13:00 crc kubenswrapper[4848]: I1204 14:13:00.675132 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc6cc5676-4j9jn" event={"ID":"9ad8065e-d54e-469b-bb54-f3a0ab26edf6","Type":"ContainerDied","Data":"5c95a5666dc776ffb5d5e6b3b266755d63c44b32ccbb3adf8b9502998c49c994"} Dec 04 14:13:01 crc kubenswrapper[4848]: I1204 14:13:01.691848 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0893eec-f2cb-41b8-a301-005dbf326afe","Type":"ContainerStarted","Data":"d0b16c6ee6e8a5c96ff397423381c1af1331b4d8345da445ec28105dfdc2da82"} Dec 04 14:13:01 crc kubenswrapper[4848]: I1204 14:13:01.940016 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6fb4f55966-dj9bq" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.196:9311/healthcheck\": read tcp 10.217.0.2:39944->10.217.0.196:9311: read: connection reset by peer" Dec 04 14:13:01 crc kubenswrapper[4848]: I1204 14:13:01.940034 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6fb4f55966-dj9bq" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.196:9311/healthcheck\": read tcp 10.217.0.2:39946->10.217.0.196:9311: read: connection reset by peer" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.562021 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.695465 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data\") pod \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.695551 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-combined-ca-bundle\") pod \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.695590 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa785c8-735e-42ff-a2dc-29481f49c4a3-logs\") pod \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.695794 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data-custom\") pod \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.695935 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9bjb\" (UniqueName: \"kubernetes.io/projected/9aa785c8-735e-42ff-a2dc-29481f49c4a3-kube-api-access-z9bjb\") pod \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\" (UID: \"9aa785c8-735e-42ff-a2dc-29481f49c4a3\") " Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.696029 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aa785c8-735e-42ff-a2dc-29481f49c4a3-logs" (OuterVolumeSpecName: "logs") pod "9aa785c8-735e-42ff-a2dc-29481f49c4a3" (UID: "9aa785c8-735e-42ff-a2dc-29481f49c4a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.696623 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9aa785c8-735e-42ff-a2dc-29481f49c4a3-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.700751 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa785c8-735e-42ff-a2dc-29481f49c4a3-kube-api-access-z9bjb" (OuterVolumeSpecName: "kube-api-access-z9bjb") pod "9aa785c8-735e-42ff-a2dc-29481f49c4a3" (UID: "9aa785c8-735e-42ff-a2dc-29481f49c4a3"). InnerVolumeSpecName "kube-api-access-z9bjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.704142 4848 generic.go:334] "Generic (PLEG): container finished" podID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerID="5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f" exitCode=0 Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.704153 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9aa785c8-735e-42ff-a2dc-29481f49c4a3" (UID: "9aa785c8-735e-42ff-a2dc-29481f49c4a3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.704194 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fb4f55966-dj9bq" event={"ID":"9aa785c8-735e-42ff-a2dc-29481f49c4a3","Type":"ContainerDied","Data":"5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f"} Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.704222 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fb4f55966-dj9bq" event={"ID":"9aa785c8-735e-42ff-a2dc-29481f49c4a3","Type":"ContainerDied","Data":"83e2a08bf38292daa3da61eb077ebe2530ef00dcc60a708c39668aedddfa6553"} Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.704240 4848 scope.go:117] "RemoveContainer" containerID="5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.704316 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fb4f55966-dj9bq" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.707709 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0893eec-f2cb-41b8-a301-005dbf326afe","Type":"ContainerStarted","Data":"badfb9e87b9641e1db1d76b3b2827de0b497bad98bc3149f1ac7b20f8c6b8824"} Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.707837 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.710966 4848 generic.go:334] "Generic (PLEG): container finished" podID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerID="2c5301185af0319c217c4f0b837fa82d49f72885cc8d9695ba56357b0224bda3" exitCode=0 Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.710993 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc6cc5676-4j9jn" event={"ID":"9ad8065e-d54e-469b-bb54-f3a0ab26edf6","Type":"ContainerDied","Data":"2c5301185af0319c217c4f0b837fa82d49f72885cc8d9695ba56357b0224bda3"} Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.763736 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.7637097539999997 podStartE2EDuration="3.763709754s" podCreationTimestamp="2025-12-04 14:12:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:02.728266594 +0000 UTC m=+1486.670763132" watchObservedRunningTime="2025-12-04 14:13:02.763709754 +0000 UTC m=+1486.706206282" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.800218 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9bjb\" (UniqueName: \"kubernetes.io/projected/9aa785c8-735e-42ff-a2dc-29481f49c4a3-kube-api-access-z9bjb\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.800450 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.801022 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9aa785c8-735e-42ff-a2dc-29481f49c4a3" (UID: "9aa785c8-735e-42ff-a2dc-29481f49c4a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.825365 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data" (OuterVolumeSpecName: "config-data") pod "9aa785c8-735e-42ff-a2dc-29481f49c4a3" (UID: "9aa785c8-735e-42ff-a2dc-29481f49c4a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.868503 4848 scope.go:117] "RemoveContainer" containerID="dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.883467 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.905309 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.905342 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa785c8-735e-42ff-a2dc-29481f49c4a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.932152 4848 scope.go:117] "RemoveContainer" containerID="5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f" Dec 04 14:13:02 crc kubenswrapper[4848]: E1204 14:13:02.935021 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f\": container with ID starting with 5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f not found: ID does not exist" containerID="5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.935083 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f"} err="failed to get container status \"5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f\": rpc error: code = NotFound desc = could not find container \"5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f\": container with ID starting with 5dfcfe54e82b6d2070478fc73cf69faa89921105d64c7491affdb81098d1c56f not found: ID does not exist" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.935135 4848 scope.go:117] "RemoveContainer" containerID="dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1" Dec 04 14:13:02 crc kubenswrapper[4848]: E1204 14:13:02.937224 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1\": container with ID starting with dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1 not found: ID does not exist" containerID="dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1" Dec 04 14:13:02 crc kubenswrapper[4848]: I1204 14:13:02.937268 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1"} err="failed to get container status \"dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1\": rpc error: code = NotFound desc = could not find container \"dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1\": container with ID starting with dc338e78b6350d671b70ed60160df05c728af57569d3dcf2c9658612a15c3bd1 not found: ID does not exist" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.006484 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qht5p\" (UniqueName: \"kubernetes.io/projected/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-kube-api-access-qht5p\") pod \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.006560 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-httpd-config\") pod \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.006623 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-ovndb-tls-certs\") pod \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.006647 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-config\") pod \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.006715 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-combined-ca-bundle\") pod \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\" (UID: \"9ad8065e-d54e-469b-bb54-f3a0ab26edf6\") " Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.011096 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-kube-api-access-qht5p" (OuterVolumeSpecName: "kube-api-access-qht5p") pod "9ad8065e-d54e-469b-bb54-f3a0ab26edf6" (UID: "9ad8065e-d54e-469b-bb54-f3a0ab26edf6"). InnerVolumeSpecName "kube-api-access-qht5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.012960 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9ad8065e-d54e-469b-bb54-f3a0ab26edf6" (UID: "9ad8065e-d54e-469b-bb54-f3a0ab26edf6"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.041154 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6fb4f55966-dj9bq"] Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.051376 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6fb4f55966-dj9bq"] Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.072540 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-config" (OuterVolumeSpecName: "config") pod "9ad8065e-d54e-469b-bb54-f3a0ab26edf6" (UID: "9ad8065e-d54e-469b-bb54-f3a0ab26edf6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.074485 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ad8065e-d54e-469b-bb54-f3a0ab26edf6" (UID: "9ad8065e-d54e-469b-bb54-f3a0ab26edf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.111198 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qht5p\" (UniqueName: \"kubernetes.io/projected/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-kube-api-access-qht5p\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.111236 4848 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.111249 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.111260 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.122577 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9ad8065e-d54e-469b-bb54-f3a0ab26edf6" (UID: "9ad8065e-d54e-469b-bb54-f3a0ab26edf6"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.153655 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k5s2p"] Dec 04 14:13:03 crc kubenswrapper[4848]: E1204 14:13:03.154172 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerName="neutron-httpd" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.154197 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerName="neutron-httpd" Dec 04 14:13:03 crc kubenswrapper[4848]: E1204 14:13:03.154218 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.154226 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api" Dec 04 14:13:03 crc kubenswrapper[4848]: E1204 14:13:03.154272 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api-log" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.154281 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api-log" Dec 04 14:13:03 crc kubenswrapper[4848]: E1204 14:13:03.154302 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerName="neutron-api" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.154313 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerName="neutron-api" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.154566 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerName="neutron-api" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.154594 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" containerName="neutron-httpd" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.154605 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.154634 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" containerName="barbican-api-log" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.156780 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.184093 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k5s2p"] Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.216477 4848 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ad8065e-d54e-469b-bb54-f3a0ab26edf6-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.318497 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5khc6\" (UniqueName: \"kubernetes.io/projected/7a22994b-5397-4d0d-9d8a-0292eca01fb4-kube-api-access-5khc6\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.318545 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-catalog-content\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.318572 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-utilities\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.421242 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5khc6\" (UniqueName: \"kubernetes.io/projected/7a22994b-5397-4d0d-9d8a-0292eca01fb4-kube-api-access-5khc6\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.421282 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-catalog-content\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.421306 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-utilities\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.421921 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-utilities\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.421920 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-catalog-content\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.443253 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5khc6\" (UniqueName: \"kubernetes.io/projected/7a22994b-5397-4d0d-9d8a-0292eca01fb4-kube-api-access-5khc6\") pod \"redhat-operators-k5s2p\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.519640 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.736929 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cc6cc5676-4j9jn" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.736980 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cc6cc5676-4j9jn" event={"ID":"9ad8065e-d54e-469b-bb54-f3a0ab26edf6","Type":"ContainerDied","Data":"a2215a2024a3cb6bc1c5957115d2c6565d7e32c1fdcd3ea79fb38b2bd7f46dbf"} Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.737045 4848 scope.go:117] "RemoveContainer" containerID="5c95a5666dc776ffb5d5e6b3b266755d63c44b32ccbb3adf8b9502998c49c994" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.765450 4848 scope.go:117] "RemoveContainer" containerID="2c5301185af0319c217c4f0b837fa82d49f72885cc8d9695ba56357b0224bda3" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.830525 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cc6cc5676-4j9jn"] Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.852428 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6cc6cc5676-4j9jn"] Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.890975 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 14:13:03 crc kubenswrapper[4848]: I1204 14:13:03.959855 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.009678 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k5s2p"] Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.027124 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pk46n"] Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.027353 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" podUID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" containerName="dnsmasq-dns" containerID="cri-o://d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56" gracePeriod=10 Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.167031 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.409446 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa785c8-735e-42ff-a2dc-29481f49c4a3" path="/var/lib/kubelet/pods/9aa785c8-735e-42ff-a2dc-29481f49c4a3/volumes" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.410488 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ad8065e-d54e-469b-bb54-f3a0ab26edf6" path="/var/lib/kubelet/pods/9ad8065e-d54e-469b-bb54-f3a0ab26edf6/volumes" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.702129 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.761466 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-nb\") pod \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.761593 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-svc\") pod \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.761694 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-sb\") pod \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.761844 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-swift-storage-0\") pod \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.761891 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-config\") pod \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.761975 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdp2t\" (UniqueName: \"kubernetes.io/projected/bb3fe94e-7191-41a1-91a8-bd35ebd781de-kube-api-access-mdp2t\") pod \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\" (UID: \"bb3fe94e-7191-41a1-91a8-bd35ebd781de\") " Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.772504 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb3fe94e-7191-41a1-91a8-bd35ebd781de-kube-api-access-mdp2t" (OuterVolumeSpecName: "kube-api-access-mdp2t") pod "bb3fe94e-7191-41a1-91a8-bd35ebd781de" (UID: "bb3fe94e-7191-41a1-91a8-bd35ebd781de"). InnerVolumeSpecName "kube-api-access-mdp2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.796383 4848 generic.go:334] "Generic (PLEG): container finished" podID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" containerID="d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56" exitCode=0 Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.796549 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.797513 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" event={"ID":"bb3fe94e-7191-41a1-91a8-bd35ebd781de","Type":"ContainerDied","Data":"d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56"} Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.797550 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-pk46n" event={"ID":"bb3fe94e-7191-41a1-91a8-bd35ebd781de","Type":"ContainerDied","Data":"ee8c337807c8db36393f21a3e7f7f965524bfc8b311e3f0194e1262eac091d85"} Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.797569 4848 scope.go:117] "RemoveContainer" containerID="d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.801590 4848 generic.go:334] "Generic (PLEG): container finished" podID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerID="d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d" exitCode=0 Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.802909 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5s2p" event={"ID":"7a22994b-5397-4d0d-9d8a-0292eca01fb4","Type":"ContainerDied","Data":"d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d"} Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.802945 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5s2p" event={"ID":"7a22994b-5397-4d0d-9d8a-0292eca01fb4","Type":"ContainerStarted","Data":"b2ac9307b7345028fb9f64be250e6d2149bdba4bd4cf4b9e1a2d2f78198173d1"} Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.818251 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb3fe94e-7191-41a1-91a8-bd35ebd781de" (UID: "bb3fe94e-7191-41a1-91a8-bd35ebd781de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.828155 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb3fe94e-7191-41a1-91a8-bd35ebd781de" (UID: "bb3fe94e-7191-41a1-91a8-bd35ebd781de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.845237 4848 scope.go:117] "RemoveContainer" containerID="4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.850212 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb3fe94e-7191-41a1-91a8-bd35ebd781de" (UID: "bb3fe94e-7191-41a1-91a8-bd35ebd781de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.868848 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.868888 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdp2t\" (UniqueName: \"kubernetes.io/projected/bb3fe94e-7191-41a1-91a8-bd35ebd781de-kube-api-access-mdp2t\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.868902 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.868914 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.897416 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb3fe94e-7191-41a1-91a8-bd35ebd781de" (UID: "bb3fe94e-7191-41a1-91a8-bd35ebd781de"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.915187 4848 scope.go:117] "RemoveContainer" containerID="d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56" Dec 04 14:13:04 crc kubenswrapper[4848]: E1204 14:13:04.917120 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56\": container with ID starting with d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56 not found: ID does not exist" containerID="d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.917183 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56"} err="failed to get container status \"d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56\": rpc error: code = NotFound desc = could not find container \"d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56\": container with ID starting with d0d93aca0ca2cb4f45be4857609a6e3d7795e6259c14708bb2f09a607b34ca56 not found: ID does not exist" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.917219 4848 scope.go:117] "RemoveContainer" containerID="4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8" Dec 04 14:13:04 crc kubenswrapper[4848]: E1204 14:13:04.920342 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8\": container with ID starting with 4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8 not found: ID does not exist" containerID="4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.920394 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8"} err="failed to get container status \"4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8\": rpc error: code = NotFound desc = could not find container \"4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8\": container with ID starting with 4d844b77914c858794605a9f725380b060e1e8fb9a3ac447d92dba67bcf000a8 not found: ID does not exist" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.928633 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-config" (OuterVolumeSpecName: "config") pod "bb3fe94e-7191-41a1-91a8-bd35ebd781de" (UID: "bb3fe94e-7191-41a1-91a8-bd35ebd781de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.929743 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.986401 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:04 crc kubenswrapper[4848]: I1204 14:13:04.986430 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb3fe94e-7191-41a1-91a8-bd35ebd781de-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:05 crc kubenswrapper[4848]: I1204 14:13:05.134847 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pk46n"] Dec 04 14:13:05 crc kubenswrapper[4848]: I1204 14:13:05.147443 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-pk46n"] Dec 04 14:13:05 crc kubenswrapper[4848]: I1204 14:13:05.816444 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5s2p" event={"ID":"7a22994b-5397-4d0d-9d8a-0292eca01fb4","Type":"ContainerStarted","Data":"2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9"} Dec 04 14:13:05 crc kubenswrapper[4848]: I1204 14:13:05.816582 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" containerName="cinder-scheduler" containerID="cri-o://e7ef746edbcd9791c05efb16f21feffebe8bd9e376a7c1fc8bac7d63f8a306fb" gracePeriod=30 Dec 04 14:13:05 crc kubenswrapper[4848]: I1204 14:13:05.816649 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" containerName="probe" containerID="cri-o://674a27ff95dcea4514c7e41e2c48a5b561c747cdd6f5d2e6fad57181047d1deb" gracePeriod=30 Dec 04 14:13:06 crc kubenswrapper[4848]: I1204 14:13:06.415496 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" path="/var/lib/kubelet/pods/bb3fe94e-7191-41a1-91a8-bd35ebd781de/volumes" Dec 04 14:13:06 crc kubenswrapper[4848]: I1204 14:13:06.838443 4848 generic.go:334] "Generic (PLEG): container finished" podID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerID="2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9" exitCode=0 Dec 04 14:13:06 crc kubenswrapper[4848]: I1204 14:13:06.839016 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5s2p" event={"ID":"7a22994b-5397-4d0d-9d8a-0292eca01fb4","Type":"ContainerDied","Data":"2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9"} Dec 04 14:13:06 crc kubenswrapper[4848]: I1204 14:13:06.843372 4848 generic.go:334] "Generic (PLEG): container finished" podID="9e19d950-51e2-4e37-9734-30330e4af07d" containerID="674a27ff95dcea4514c7e41e2c48a5b561c747cdd6f5d2e6fad57181047d1deb" exitCode=0 Dec 04 14:13:06 crc kubenswrapper[4848]: I1204 14:13:06.843475 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e19d950-51e2-4e37-9734-30330e4af07d","Type":"ContainerDied","Data":"674a27ff95dcea4514c7e41e2c48a5b561c747cdd6f5d2e6fad57181047d1deb"} Dec 04 14:13:08 crc kubenswrapper[4848]: I1204 14:13:08.871050 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5s2p" event={"ID":"7a22994b-5397-4d0d-9d8a-0292eca01fb4","Type":"ContainerStarted","Data":"3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a"} Dec 04 14:13:08 crc kubenswrapper[4848]: I1204 14:13:08.895034 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k5s2p" podStartSLOduration=2.865837342 podStartE2EDuration="5.895015093s" podCreationTimestamp="2025-12-04 14:13:03 +0000 UTC" firstStartedPulling="2025-12-04 14:13:04.81808121 +0000 UTC m=+1488.760577738" lastFinishedPulling="2025-12-04 14:13:07.847258961 +0000 UTC m=+1491.789755489" observedRunningTime="2025-12-04 14:13:08.889051329 +0000 UTC m=+1492.831547867" watchObservedRunningTime="2025-12-04 14:13:08.895015093 +0000 UTC m=+1492.837511631" Dec 04 14:13:09 crc kubenswrapper[4848]: I1204 14:13:09.666532 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5b74ff8c74-nm59b" Dec 04 14:13:09 crc kubenswrapper[4848]: I1204 14:13:09.889199 4848 generic.go:334] "Generic (PLEG): container finished" podID="9e19d950-51e2-4e37-9734-30330e4af07d" containerID="e7ef746edbcd9791c05efb16f21feffebe8bd9e376a7c1fc8bac7d63f8a306fb" exitCode=0 Dec 04 14:13:09 crc kubenswrapper[4848]: I1204 14:13:09.889308 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e19d950-51e2-4e37-9734-30330e4af07d","Type":"ContainerDied","Data":"e7ef746edbcd9791c05efb16f21feffebe8bd9e376a7c1fc8bac7d63f8a306fb"} Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.154066 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.202311 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data-custom\") pod \"9e19d950-51e2-4e37-9734-30330e4af07d\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.202348 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e19d950-51e2-4e37-9734-30330e4af07d-etc-machine-id\") pod \"9e19d950-51e2-4e37-9734-30330e4af07d\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.202438 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e19d950-51e2-4e37-9734-30330e4af07d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9e19d950-51e2-4e37-9734-30330e4af07d" (UID: "9e19d950-51e2-4e37-9734-30330e4af07d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.202559 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data\") pod \"9e19d950-51e2-4e37-9734-30330e4af07d\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.202590 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-scripts\") pod \"9e19d950-51e2-4e37-9734-30330e4af07d\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.202610 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-combined-ca-bundle\") pod \"9e19d950-51e2-4e37-9734-30330e4af07d\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.202662 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f72wc\" (UniqueName: \"kubernetes.io/projected/9e19d950-51e2-4e37-9734-30330e4af07d-kube-api-access-f72wc\") pod \"9e19d950-51e2-4e37-9734-30330e4af07d\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.203115 4848 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e19d950-51e2-4e37-9734-30330e4af07d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.209465 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e19d950-51e2-4e37-9734-30330e4af07d-kube-api-access-f72wc" (OuterVolumeSpecName: "kube-api-access-f72wc") pod "9e19d950-51e2-4e37-9734-30330e4af07d" (UID: "9e19d950-51e2-4e37-9734-30330e4af07d"). InnerVolumeSpecName "kube-api-access-f72wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.212455 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9e19d950-51e2-4e37-9734-30330e4af07d" (UID: "9e19d950-51e2-4e37-9734-30330e4af07d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.232217 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-scripts" (OuterVolumeSpecName: "scripts") pod "9e19d950-51e2-4e37-9734-30330e4af07d" (UID: "9e19d950-51e2-4e37-9734-30330e4af07d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.305998 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.306366 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.306443 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f72wc\" (UniqueName: \"kubernetes.io/projected/9e19d950-51e2-4e37-9734-30330e4af07d-kube-api-access-f72wc\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.307643 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e19d950-51e2-4e37-9734-30330e4af07d" (UID: "9e19d950-51e2-4e37-9734-30330e4af07d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.407126 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data" (OuterVolumeSpecName: "config-data") pod "9e19d950-51e2-4e37-9734-30330e4af07d" (UID: "9e19d950-51e2-4e37-9734-30330e4af07d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.407428 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data\") pod \"9e19d950-51e2-4e37-9734-30330e4af07d\" (UID: \"9e19d950-51e2-4e37-9734-30330e4af07d\") " Dec 04 14:13:10 crc kubenswrapper[4848]: W1204 14:13:10.407555 4848 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/9e19d950-51e2-4e37-9734-30330e4af07d/volumes/kubernetes.io~secret/config-data Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.407565 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data" (OuterVolumeSpecName: "config-data") pod "9e19d950-51e2-4e37-9734-30330e4af07d" (UID: "9e19d950-51e2-4e37-9734-30330e4af07d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.408070 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.408084 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e19d950-51e2-4e37-9734-30330e4af07d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.909161 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e19d950-51e2-4e37-9734-30330e4af07d","Type":"ContainerDied","Data":"d542ab438d63d7c7a94d7a7aa80a1b2c38a25af0a31a271aca659e64ba71f6e1"} Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.909218 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.909233 4848 scope.go:117] "RemoveContainer" containerID="674a27ff95dcea4514c7e41e2c48a5b561c747cdd6f5d2e6fad57181047d1deb" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.951139 4848 scope.go:117] "RemoveContainer" containerID="e7ef746edbcd9791c05efb16f21feffebe8bd9e376a7c1fc8bac7d63f8a306fb" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.952906 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.963882 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.991178 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:13:10 crc kubenswrapper[4848]: E1204 14:13:10.991723 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" containerName="dnsmasq-dns" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.991742 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" containerName="dnsmasq-dns" Dec 04 14:13:10 crc kubenswrapper[4848]: E1204 14:13:10.991754 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" containerName="cinder-scheduler" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.991760 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" containerName="cinder-scheduler" Dec 04 14:13:10 crc kubenswrapper[4848]: E1204 14:13:10.991778 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" containerName="probe" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.991785 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" containerName="probe" Dec 04 14:13:10 crc kubenswrapper[4848]: E1204 14:13:10.991797 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" containerName="init" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.991803 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" containerName="init" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.992245 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" containerName="probe" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.992283 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb3fe94e-7191-41a1-91a8-bd35ebd781de" containerName="dnsmasq-dns" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.992297 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" containerName="cinder-scheduler" Dec 04 14:13:10 crc kubenswrapper[4848]: I1204 14:13:10.993565 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.022294 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.029572 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.124170 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmt9l\" (UniqueName: \"kubernetes.io/projected/c93bea85-4d93-4003-b146-70a07e1edd92-kube-api-access-lmt9l\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.124292 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.124376 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-scripts\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.124399 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.124526 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-config-data\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.124555 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c93bea85-4d93-4003-b146-70a07e1edd92-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.226777 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmt9l\" (UniqueName: \"kubernetes.io/projected/c93bea85-4d93-4003-b146-70a07e1edd92-kube-api-access-lmt9l\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.227191 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.227267 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-scripts\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.227291 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.227415 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-config-data\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.227446 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c93bea85-4d93-4003-b146-70a07e1edd92-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.227539 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c93bea85-4d93-4003-b146-70a07e1edd92-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.231248 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-scripts\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.232063 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.235753 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.237821 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c93bea85-4d93-4003-b146-70a07e1edd92-config-data\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.245766 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmt9l\" (UniqueName: \"kubernetes.io/projected/c93bea85-4d93-4003-b146-70a07e1edd92-kube-api-access-lmt9l\") pod \"cinder-scheduler-0\" (UID: \"c93bea85-4d93-4003-b146-70a07e1edd92\") " pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.341787 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.841179 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 14:13:11 crc kubenswrapper[4848]: I1204 14:13:11.924021 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c93bea85-4d93-4003-b146-70a07e1edd92","Type":"ContainerStarted","Data":"14c742e36461819fe6adf302c1873938215ef59e83b00dd537a483cbbc6c35bb"} Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.185757 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.187736 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.195062 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.195104 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-bg4xp" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.195448 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.207380 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.259261 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec3f1251-eb61-4b0e-b534-177c7d928404-openstack-config-secret\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.259465 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec3f1251-eb61-4b0e-b534-177c7d928404-openstack-config\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.259575 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb4tj\" (UniqueName: \"kubernetes.io/projected/ec3f1251-eb61-4b0e-b534-177c7d928404-kube-api-access-rb4tj\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.259688 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec3f1251-eb61-4b0e-b534-177c7d928404-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.361391 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec3f1251-eb61-4b0e-b534-177c7d928404-openstack-config-secret\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.361527 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec3f1251-eb61-4b0e-b534-177c7d928404-openstack-config\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.361620 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb4tj\" (UniqueName: \"kubernetes.io/projected/ec3f1251-eb61-4b0e-b534-177c7d928404-kube-api-access-rb4tj\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.361696 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec3f1251-eb61-4b0e-b534-177c7d928404-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.372025 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec3f1251-eb61-4b0e-b534-177c7d928404-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.373290 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec3f1251-eb61-4b0e-b534-177c7d928404-openstack-config\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.375272 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec3f1251-eb61-4b0e-b534-177c7d928404-openstack-config-secret\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.385182 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb4tj\" (UniqueName: \"kubernetes.io/projected/ec3f1251-eb61-4b0e-b534-177c7d928404-kube-api-access-rb4tj\") pod \"openstackclient\" (UID: \"ec3f1251-eb61-4b0e-b534-177c7d928404\") " pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.407079 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e19d950-51e2-4e37-9734-30330e4af07d" path="/var/lib/kubelet/pods/9e19d950-51e2-4e37-9734-30330e4af07d/volumes" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.515334 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 14:13:12 crc kubenswrapper[4848]: I1204 14:13:12.613533 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 14:13:13 crc kubenswrapper[4848]: I1204 14:13:13.004978 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 14:13:13 crc kubenswrapper[4848]: I1204 14:13:13.522191 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:13 crc kubenswrapper[4848]: I1204 14:13:13.522255 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:13 crc kubenswrapper[4848]: I1204 14:13:13.946730 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ec3f1251-eb61-4b0e-b534-177c7d928404","Type":"ContainerStarted","Data":"60f8490d39d00ba97767ce9847c186469da03633e56ac7eae03724ace6d12ab8"} Dec 04 14:13:13 crc kubenswrapper[4848]: I1204 14:13:13.949777 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c93bea85-4d93-4003-b146-70a07e1edd92","Type":"ContainerStarted","Data":"749e00937f4eb887aa1920c2f877f2344e71a220f398084909bc1b96675d0bf7"} Dec 04 14:13:14 crc kubenswrapper[4848]: I1204 14:13:14.573615 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k5s2p" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="registry-server" probeResult="failure" output=< Dec 04 14:13:14 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 14:13:14 crc kubenswrapper[4848]: > Dec 04 14:13:14 crc kubenswrapper[4848]: I1204 14:13:14.966340 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c93bea85-4d93-4003-b146-70a07e1edd92","Type":"ContainerStarted","Data":"d6beb1b9d944099e7df40e557bb37bbc5f7ddf2a2460fa3c30ff0637d8d309f9"} Dec 04 14:13:14 crc kubenswrapper[4848]: I1204 14:13:14.986367 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.986349727 podStartE2EDuration="4.986349727s" podCreationTimestamp="2025-12-04 14:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:14.982500099 +0000 UTC m=+1498.924996627" watchObservedRunningTime="2025-12-04 14:13:14.986349727 +0000 UTC m=+1498.928846255" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.343431 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.826781 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-b65c5f4cf-q22f8"] Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.831763 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.854077 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-b65c5f4cf-q22f8"] Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.863241 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.863505 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.864040 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.886885 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-config-data\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.886931 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-combined-ca-bundle\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.886988 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acf82280-f41a-460a-84ab-639c541670fc-log-httpd\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.887009 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/acf82280-f41a-460a-84ab-639c541670fc-etc-swift\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.887068 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djlgz\" (UniqueName: \"kubernetes.io/projected/acf82280-f41a-460a-84ab-639c541670fc-kube-api-access-djlgz\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.887108 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-internal-tls-certs\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.887163 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acf82280-f41a-460a-84ab-639c541670fc-run-httpd\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.887235 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-public-tls-certs\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.992432 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-public-tls-certs\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.992575 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-config-data\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.992615 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-combined-ca-bundle\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.992684 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/acf82280-f41a-460a-84ab-639c541670fc-etc-swift\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.992703 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acf82280-f41a-460a-84ab-639c541670fc-log-httpd\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.992806 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djlgz\" (UniqueName: \"kubernetes.io/projected/acf82280-f41a-460a-84ab-639c541670fc-kube-api-access-djlgz\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.992890 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-internal-tls-certs\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.992992 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acf82280-f41a-460a-84ab-639c541670fc-run-httpd\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:16 crc kubenswrapper[4848]: I1204 14:13:16.993798 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acf82280-f41a-460a-84ab-639c541670fc-run-httpd\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:16.999557 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/acf82280-f41a-460a-84ab-639c541670fc-log-httpd\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:17.022675 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-config-data\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:17.023177 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-public-tls-certs\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:17.051223 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-combined-ca-bundle\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:17.052093 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djlgz\" (UniqueName: \"kubernetes.io/projected/acf82280-f41a-460a-84ab-639c541670fc-kube-api-access-djlgz\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:17.055167 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/acf82280-f41a-460a-84ab-639c541670fc-etc-swift\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:17.055941 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acf82280-f41a-460a-84ab-639c541670fc-internal-tls-certs\") pod \"swift-proxy-b65c5f4cf-q22f8\" (UID: \"acf82280-f41a-460a-84ab-639c541670fc\") " pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:17.170429 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:17 crc kubenswrapper[4848]: I1204 14:13:17.923536 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-b65c5f4cf-q22f8"] Dec 04 14:13:18 crc kubenswrapper[4848]: I1204 14:13:18.482112 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8766cc68-d2824" Dec 04 14:13:18 crc kubenswrapper[4848]: I1204 14:13:18.488310 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8766cc68-d2824" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.634053 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7467697848-tlr74"] Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.636017 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.641687 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-bhc6n" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.641885 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.642014 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.644536 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7467697848-tlr74"] Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.779114 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data-custom\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.783042 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2srvl\" (UniqueName: \"kubernetes.io/projected/ca5459c4-87fe-414e-9e6e-843d0e6329a9-kube-api-access-2srvl\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.783357 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-combined-ca-bundle\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.787165 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.806315 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-fhzgf"] Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.819095 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.860473 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-fhzgf"] Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.874105 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-65b648c688-s9c2l"] Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.876503 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.886458 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.886695 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-665fbf8d68-zvxk9"] Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.889261 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.891090 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data-custom\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.891133 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2srvl\" (UniqueName: \"kubernetes.io/projected/ca5459c4-87fe-414e-9e6e-843d0e6329a9-kube-api-access-2srvl\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.891165 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.891262 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-combined-ca-bundle\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.891313 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.897936 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-65b648c688-s9c2l"] Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.901518 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.901663 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-combined-ca-bundle\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.913278 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data-custom\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.916083 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-665fbf8d68-zvxk9"] Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.924474 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2srvl\" (UniqueName: \"kubernetes.io/projected/ca5459c4-87fe-414e-9e6e-843d0e6329a9-kube-api-access-2srvl\") pod \"heat-engine-7467697848-tlr74\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.971468 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995022 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-combined-ca-bundle\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995101 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995162 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kwpz\" (UniqueName: \"kubernetes.io/projected/70bfcde6-2ad2-44fc-ae39-5821c63f4918-kube-api-access-4kwpz\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995192 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data-custom\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995272 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-combined-ca-bundle\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995349 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995405 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-config\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995425 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995444 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995772 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995825 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpvbv\" (UniqueName: \"kubernetes.io/projected/b38b0b7c-d339-418d-b2c4-0278a451626a-kube-api-access-cpvbv\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995854 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995869 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8twp\" (UniqueName: \"kubernetes.io/projected/d58a4892-6cad-4459-b103-77bdbba3e42a-kube-api-access-j8twp\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:19 crc kubenswrapper[4848]: I1204 14:13:19.995886 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data-custom\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098147 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098225 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpvbv\" (UniqueName: \"kubernetes.io/projected/b38b0b7c-d339-418d-b2c4-0278a451626a-kube-api-access-cpvbv\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098255 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098271 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8twp\" (UniqueName: \"kubernetes.io/projected/d58a4892-6cad-4459-b103-77bdbba3e42a-kube-api-access-j8twp\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098288 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data-custom\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098363 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-combined-ca-bundle\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098382 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098403 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kwpz\" (UniqueName: \"kubernetes.io/projected/70bfcde6-2ad2-44fc-ae39-5821c63f4918-kube-api-access-4kwpz\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098427 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data-custom\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098488 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-combined-ca-bundle\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098516 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098544 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-config\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098561 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.098579 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.099775 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.101114 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-config\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.101610 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.101906 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.102621 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.105531 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.106496 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data-custom\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.108634 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data-custom\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.110045 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.110215 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-combined-ca-bundle\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.112879 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-combined-ca-bundle\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.120139 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8twp\" (UniqueName: \"kubernetes.io/projected/d58a4892-6cad-4459-b103-77bdbba3e42a-kube-api-access-j8twp\") pod \"dnsmasq-dns-7756b9d78c-fhzgf\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.126411 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kwpz\" (UniqueName: \"kubernetes.io/projected/70bfcde6-2ad2-44fc-ae39-5821c63f4918-kube-api-access-4kwpz\") pod \"heat-api-65b648c688-s9c2l\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.126761 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpvbv\" (UniqueName: \"kubernetes.io/projected/b38b0b7c-d339-418d-b2c4-0278a451626a-kube-api-access-cpvbv\") pod \"heat-cfnapi-665fbf8d68-zvxk9\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.178215 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.316292 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:20 crc kubenswrapper[4848]: I1204 14:13:20.324077 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:21 crc kubenswrapper[4848]: I1204 14:13:21.023591 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="99e49830-7634-4257-af28-938f373af926" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 04 14:13:21 crc kubenswrapper[4848]: I1204 14:13:21.599629 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 14:13:23 crc kubenswrapper[4848]: I1204 14:13:23.581026 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:23 crc kubenswrapper[4848]: I1204 14:13:23.635040 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:23 crc kubenswrapper[4848]: I1204 14:13:23.828159 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k5s2p"] Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.212580 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ec3f1251-eb61-4b0e-b534-177c7d928404","Type":"ContainerStarted","Data":"8b26ec4f9d6dc21cf04ebc63fa55d7bd022afad3d771b6bb7dea3fc474df7708"} Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.216326 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b65c5f4cf-q22f8" event={"ID":"acf82280-f41a-460a-84ab-639c541670fc","Type":"ContainerStarted","Data":"61fb3dba04f684e9df4a395214700daf43cf444a85f58a9ab4eeff5cfdf8322e"} Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.216360 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k5s2p" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="registry-server" containerID="cri-o://3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a" gracePeriod=2 Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.216385 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b65c5f4cf-q22f8" event={"ID":"acf82280-f41a-460a-84ab-639c541670fc","Type":"ContainerStarted","Data":"eeb771d2a176649c738b1c6947b6fa891f541e9c9630586b9294933a641ad868"} Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.231290 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.407721134 podStartE2EDuration="13.231270875s" podCreationTimestamp="2025-12-04 14:13:12 +0000 UTC" firstStartedPulling="2025-12-04 14:13:13.011889237 +0000 UTC m=+1496.954385765" lastFinishedPulling="2025-12-04 14:13:24.835438978 +0000 UTC m=+1508.777935506" observedRunningTime="2025-12-04 14:13:25.230474767 +0000 UTC m=+1509.172971295" watchObservedRunningTime="2025-12-04 14:13:25.231270875 +0000 UTC m=+1509.173767403" Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.409721 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-fhzgf"] Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.426771 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7467697848-tlr74"] Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.565562 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-65b648c688-s9c2l"] Dec 04 14:13:25 crc kubenswrapper[4848]: W1204 14:13:25.623187 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70bfcde6_2ad2_44fc_ae39_5821c63f4918.slice/crio-4df30f9ebbee66082ff0a6f20c3e1bcfa1b58470025e84b96a89d379b6db8cd0 WatchSource:0}: Error finding container 4df30f9ebbee66082ff0a6f20c3e1bcfa1b58470025e84b96a89d379b6db8cd0: Status 404 returned error can't find the container with id 4df30f9ebbee66082ff0a6f20c3e1bcfa1b58470025e84b96a89d379b6db8cd0 Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.679068 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-665fbf8d68-zvxk9"] Dec 04 14:13:25 crc kubenswrapper[4848]: W1204 14:13:25.681161 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb38b0b7c_d339_418d_b2c4_0278a451626a.slice/crio-534c15b09d5ca3658c48975811ff39749c6fd3c04024da2536f2f57e74ea48c9 WatchSource:0}: Error finding container 534c15b09d5ca3658c48975811ff39749c6fd3c04024da2536f2f57e74ea48c9: Status 404 returned error can't find the container with id 534c15b09d5ca3658c48975811ff39749c6fd3c04024da2536f2f57e74ea48c9 Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.866248 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.866806 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-log" containerID="cri-o://c7727796ae2e569ffb4887e1d8acaaae87939cc44b10b82cc3a84810cd2191aa" gracePeriod=30 Dec 04 14:13:25 crc kubenswrapper[4848]: I1204 14:13:25.869571 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-httpd" containerID="cri-o://099fcfe1d45f56ac6d66db57942f93cf89b334794714da0f9c2e4368156413a1" gracePeriod=30 Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.053185 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-55dd7d4b8f-9bj88"] Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.069708 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.089194 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55dd7d4b8f-9bj88"] Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.097041 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-878c6546-2c98w"] Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.098904 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.109389 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-58569cf956-f89kq"] Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.118611 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.166579 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-878c6546-2c98w"] Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.191191 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58569cf956-f89kq"] Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.233827 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scrbj\" (UniqueName: \"kubernetes.io/projected/732df5f3-a0ba-449b-bf97-8604a569314f-kube-api-access-scrbj\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.275258 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-b65c5f4cf-q22f8" event={"ID":"acf82280-f41a-460a-84ab-639c541670fc","Type":"ContainerStarted","Data":"d8d477d05c18355d82b7b93ed45a9971f0d5dc34563c8601396d93981d362b84"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.276193 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.276330 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.277026 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" event={"ID":"b38b0b7c-d339-418d-b2c4-0278a451626a","Type":"ContainerStarted","Data":"534c15b09d5ca3658c48975811ff39749c6fd3c04024da2536f2f57e74ea48c9"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.279751 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data-custom\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.279865 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/28dda188-1082-4b1b-a0dc-42297a0918b4-kube-api-access-cdcsz\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.279918 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4tj9\" (UniqueName: \"kubernetes.io/projected/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-kube-api-access-g4tj9\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.280021 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data-custom\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.280078 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.280148 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data-custom\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.280239 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-combined-ca-bundle\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.280435 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.280494 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.280513 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-combined-ca-bundle\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.280571 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-combined-ca-bundle\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.282905 4848 generic.go:334] "Generic (PLEG): container finished" podID="d58a4892-6cad-4459-b103-77bdbba3e42a" containerID="e29d436c8560a47c9e20661f65c6536970d5a6f67a0d5b31ec521f1c83361995" exitCode=0 Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.282995 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" event={"ID":"d58a4892-6cad-4459-b103-77bdbba3e42a","Type":"ContainerDied","Data":"e29d436c8560a47c9e20661f65c6536970d5a6f67a0d5b31ec521f1c83361995"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.283021 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" event={"ID":"d58a4892-6cad-4459-b103-77bdbba3e42a","Type":"ContainerStarted","Data":"d1498f5a4a349cf0a8105841a23fb6ac74fe277e758fe841de52bbd0017cfa6f"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.285059 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.290084 4848 generic.go:334] "Generic (PLEG): container finished" podID="99e49830-7634-4257-af28-938f373af926" containerID="615184e23c8aeb0ec7d6d6030f4b2dcd7cf611c05c94c19a1cf1b3569d831af2" exitCode=137 Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.290233 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99e49830-7634-4257-af28-938f373af926","Type":"ContainerDied","Data":"615184e23c8aeb0ec7d6d6030f4b2dcd7cf611c05c94c19a1cf1b3569d831af2"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.290295 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99e49830-7634-4257-af28-938f373af926","Type":"ContainerDied","Data":"9dea439f054cfaf9d67c7b11625a8be33e60f1e0416b83f1defc9451ff6186b4"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.290317 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dea439f054cfaf9d67c7b11625a8be33e60f1e0416b83f1defc9451ff6186b4" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.315568 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-b65c5f4cf-q22f8" podStartSLOduration=10.315544103 podStartE2EDuration="10.315544103s" podCreationTimestamp="2025-12-04 14:13:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:26.293485034 +0000 UTC m=+1510.235981582" watchObservedRunningTime="2025-12-04 14:13:26.315544103 +0000 UTC m=+1510.258040631" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.320484 4848 generic.go:334] "Generic (PLEG): container finished" podID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerID="3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a" exitCode=0 Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.320577 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5s2p" event={"ID":"7a22994b-5397-4d0d-9d8a-0292eca01fb4","Type":"ContainerDied","Data":"3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.320622 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k5s2p" event={"ID":"7a22994b-5397-4d0d-9d8a-0292eca01fb4","Type":"ContainerDied","Data":"b2ac9307b7345028fb9f64be250e6d2149bdba4bd4cf4b9e1a2d2f78198173d1"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.320638 4848 scope.go:117] "RemoveContainer" containerID="3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.321143 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k5s2p" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.349133 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7467697848-tlr74" event={"ID":"ca5459c4-87fe-414e-9e6e-843d0e6329a9","Type":"ContainerStarted","Data":"7558f74fc22966b329e7619df76d7df2a8dcb0d46671e858ea57020159bf533b"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.349192 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7467697848-tlr74" event={"ID":"ca5459c4-87fe-414e-9e6e-843d0e6329a9","Type":"ContainerStarted","Data":"62fa182304a6d6657dc4a2093ffb6178169940a3a8347e0cf0428624401fc5ea"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.349496 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.361525 4848 generic.go:334] "Generic (PLEG): container finished" podID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerID="c7727796ae2e569ffb4887e1d8acaaae87939cc44b10b82cc3a84810cd2191aa" exitCode=143 Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.362102 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db70a1ef-eaa8-4da0-8abe-72aa5c822226","Type":"ContainerDied","Data":"c7727796ae2e569ffb4887e1d8acaaae87939cc44b10b82cc3a84810cd2191aa"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.382219 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-utilities\") pod \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.382318 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-catalog-content\") pod \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.382448 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5khc6\" (UniqueName: \"kubernetes.io/projected/7a22994b-5397-4d0d-9d8a-0292eca01fb4-kube-api-access-5khc6\") pod \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\" (UID: \"7a22994b-5397-4d0d-9d8a-0292eca01fb4\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.383582 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-combined-ca-bundle\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.383712 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.383746 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-combined-ca-bundle\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.383763 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.383884 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-combined-ca-bundle\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.383985 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scrbj\" (UniqueName: \"kubernetes.io/projected/732df5f3-a0ba-449b-bf97-8604a569314f-kube-api-access-scrbj\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.384083 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data-custom\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.384344 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/28dda188-1082-4b1b-a0dc-42297a0918b4-kube-api-access-cdcsz\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.384381 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-utilities" (OuterVolumeSpecName: "utilities") pod "7a22994b-5397-4d0d-9d8a-0292eca01fb4" (UID: "7a22994b-5397-4d0d-9d8a-0292eca01fb4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.384404 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4tj9\" (UniqueName: \"kubernetes.io/projected/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-kube-api-access-g4tj9\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.384549 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data-custom\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.384619 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.384690 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data-custom\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.384957 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.396640 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-combined-ca-bundle\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.397432 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.397712 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.399703 4848 scope.go:117] "RemoveContainer" containerID="2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.402485 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data-custom\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.402483 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.405351 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a22994b-5397-4d0d-9d8a-0292eca01fb4-kube-api-access-5khc6" (OuterVolumeSpecName: "kube-api-access-5khc6") pod "7a22994b-5397-4d0d-9d8a-0292eca01fb4" (UID: "7a22994b-5397-4d0d-9d8a-0292eca01fb4"). InnerVolumeSpecName "kube-api-access-5khc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.406034 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4tj9\" (UniqueName: \"kubernetes.io/projected/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-kube-api-access-g4tj9\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.407395 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/28dda188-1082-4b1b-a0dc-42297a0918b4-kube-api-access-cdcsz\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.408157 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scrbj\" (UniqueName: \"kubernetes.io/projected/732df5f3-a0ba-449b-bf97-8604a569314f-kube-api-access-scrbj\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.411057 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.415409 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7467697848-tlr74" podStartSLOduration=7.415387509 podStartE2EDuration="7.415387509s" podCreationTimestamp="2025-12-04 14:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:26.370576646 +0000 UTC m=+1510.313073174" watchObservedRunningTime="2025-12-04 14:13:26.415387509 +0000 UTC m=+1510.357884037" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.416617 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data-custom\") pod \"heat-cfnapi-58569cf956-f89kq\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.417637 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-combined-ca-bundle\") pod \"heat-engine-55dd7d4b8f-9bj88\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.426331 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-combined-ca-bundle\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.463707 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data-custom\") pod \"heat-api-878c6546-2c98w\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.475604 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-65b648c688-s9c2l" event={"ID":"70bfcde6-2ad2-44fc-ae39-5821c63f4918","Type":"ContainerStarted","Data":"4df30f9ebbee66082ff0a6f20c3e1bcfa1b58470025e84b96a89d379b6db8cd0"} Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.484621 4848 scope.go:117] "RemoveContainer" containerID="d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.487094 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-scripts\") pod \"99e49830-7634-4257-af28-938f373af926\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.487246 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-config-data\") pod \"99e49830-7634-4257-af28-938f373af926\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.487301 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-combined-ca-bundle\") pod \"99e49830-7634-4257-af28-938f373af926\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.487345 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmzt6\" (UniqueName: \"kubernetes.io/projected/99e49830-7634-4257-af28-938f373af926-kube-api-access-zmzt6\") pod \"99e49830-7634-4257-af28-938f373af926\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.487392 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-sg-core-conf-yaml\") pod \"99e49830-7634-4257-af28-938f373af926\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.487421 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-run-httpd\") pod \"99e49830-7634-4257-af28-938f373af926\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.487484 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-log-httpd\") pod \"99e49830-7634-4257-af28-938f373af926\" (UID: \"99e49830-7634-4257-af28-938f373af926\") " Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.488189 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5khc6\" (UniqueName: \"kubernetes.io/projected/7a22994b-5397-4d0d-9d8a-0292eca01fb4-kube-api-access-5khc6\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.491158 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "99e49830-7634-4257-af28-938f373af926" (UID: "99e49830-7634-4257-af28-938f373af926"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.492595 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "99e49830-7634-4257-af28-938f373af926" (UID: "99e49830-7634-4257-af28-938f373af926"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.496398 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e49830-7634-4257-af28-938f373af926-kube-api-access-zmzt6" (OuterVolumeSpecName: "kube-api-access-zmzt6") pod "99e49830-7634-4257-af28-938f373af926" (UID: "99e49830-7634-4257-af28-938f373af926"). InnerVolumeSpecName "kube-api-access-zmzt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.500151 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-scripts" (OuterVolumeSpecName: "scripts") pod "99e49830-7634-4257-af28-938f373af926" (UID: "99e49830-7634-4257-af28-938f373af926"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.550136 4848 scope.go:117] "RemoveContainer" containerID="3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a" Dec 04 14:13:26 crc kubenswrapper[4848]: E1204 14:13:26.552148 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a\": container with ID starting with 3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a not found: ID does not exist" containerID="3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.552218 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a"} err="failed to get container status \"3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a\": rpc error: code = NotFound desc = could not find container \"3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a\": container with ID starting with 3e3e390102f7b6649aa65ae0cf1abd778a0da67e1b89cf39cbe334c098b5eb7a not found: ID does not exist" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.552246 4848 scope.go:117] "RemoveContainer" containerID="2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9" Dec 04 14:13:26 crc kubenswrapper[4848]: E1204 14:13:26.554309 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9\": container with ID starting with 2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9 not found: ID does not exist" containerID="2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.554388 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9"} err="failed to get container status \"2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9\": rpc error: code = NotFound desc = could not find container \"2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9\": container with ID starting with 2b87095cce731f64ba94b0b08a502cc231173d346fc12d7497e6adf8f5a806c9 not found: ID does not exist" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.554428 4848 scope.go:117] "RemoveContainer" containerID="d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d" Dec 04 14:13:26 crc kubenswrapper[4848]: E1204 14:13:26.558209 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d\": container with ID starting with d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d not found: ID does not exist" containerID="d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.558270 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d"} err="failed to get container status \"d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d\": rpc error: code = NotFound desc = could not find container \"d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d\": container with ID starting with d3c34265316d75f6f9b46df7f71f5c4c0fbe42e616fcb5da49406287403f799d not found: ID does not exist" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.574324 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.582198 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a22994b-5397-4d0d-9d8a-0292eca01fb4" (UID: "7a22994b-5397-4d0d-9d8a-0292eca01fb4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.590653 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.590682 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a22994b-5397-4d0d-9d8a-0292eca01fb4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.590692 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmzt6\" (UniqueName: \"kubernetes.io/projected/99e49830-7634-4257-af28-938f373af926-kube-api-access-zmzt6\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.590701 4848 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.590712 4848 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99e49830-7634-4257-af28-938f373af926-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.593532 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "99e49830-7634-4257-af28-938f373af926" (UID: "99e49830-7634-4257-af28-938f373af926"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.608723 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99e49830-7634-4257-af28-938f373af926" (UID: "99e49830-7634-4257-af28-938f373af926"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.625872 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.647656 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.692941 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.692987 4848 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.709702 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-config-data" (OuterVolumeSpecName: "config-data") pod "99e49830-7634-4257-af28-938f373af926" (UID: "99e49830-7634-4257-af28-938f373af926"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.794863 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99e49830-7634-4257-af28-938f373af926-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.880502 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k5s2p"] Dec 04 14:13:26 crc kubenswrapper[4848]: I1204 14:13:26.900216 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k5s2p"] Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.157017 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55dd7d4b8f-9bj88"] Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.185848 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-878c6546-2c98w"] Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.437510 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-878c6546-2c98w" event={"ID":"a9478f9e-a08b-4d95-b57a-fb8ed97850fb","Type":"ContainerStarted","Data":"d7e28a1e8e7ee8c99e9d551eeaf2e177f76306841e0b48c0be27bd0a8a8cb657"} Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.445323 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" event={"ID":"d58a4892-6cad-4459-b103-77bdbba3e42a","Type":"ContainerStarted","Data":"32e3e0bffe950f8c8455df58282141d62ad507d863dce698cd847fca738818f0"} Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.445898 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.455996 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55dd7d4b8f-9bj88" event={"ID":"732df5f3-a0ba-449b-bf97-8604a569314f","Type":"ContainerStarted","Data":"c0ec08d301c28b573439ae3ab2329665c0a216ff3bcdf8f54dbb2221b219cb08"} Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.456138 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.486553 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" podStartSLOduration=8.48653428 podStartE2EDuration="8.48653428s" podCreationTimestamp="2025-12-04 14:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:27.466532029 +0000 UTC m=+1511.409028557" watchObservedRunningTime="2025-12-04 14:13:27.48653428 +0000 UTC m=+1511.429030808" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.498881 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58569cf956-f89kq"] Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.553666 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.569771 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.582290 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:27 crc kubenswrapper[4848]: E1204 14:13:27.582931 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e49830-7634-4257-af28-938f373af926" containerName="ceilometer-notification-agent" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.582972 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e49830-7634-4257-af28-938f373af926" containerName="ceilometer-notification-agent" Dec 04 14:13:27 crc kubenswrapper[4848]: E1204 14:13:27.582992 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="extract-content" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583001 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="extract-content" Dec 04 14:13:27 crc kubenswrapper[4848]: E1204 14:13:27.583036 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e49830-7634-4257-af28-938f373af926" containerName="proxy-httpd" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583043 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e49830-7634-4257-af28-938f373af926" containerName="proxy-httpd" Dec 04 14:13:27 crc kubenswrapper[4848]: E1204 14:13:27.583065 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e49830-7634-4257-af28-938f373af926" containerName="sg-core" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583071 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e49830-7634-4257-af28-938f373af926" containerName="sg-core" Dec 04 14:13:27 crc kubenswrapper[4848]: E1204 14:13:27.583084 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="registry-server" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583092 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="registry-server" Dec 04 14:13:27 crc kubenswrapper[4848]: E1204 14:13:27.583109 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="extract-utilities" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583115 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="extract-utilities" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583307 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" containerName="registry-server" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583319 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e49830-7634-4257-af28-938f373af926" containerName="ceilometer-notification-agent" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583330 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e49830-7634-4257-af28-938f373af926" containerName="proxy-httpd" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.583353 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e49830-7634-4257-af28-938f373af926" containerName="sg-core" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.585346 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.594761 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.595051 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.623788 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.635704 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-scripts\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.636323 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.636418 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-run-httpd\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.636451 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.636523 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb248\" (UniqueName: \"kubernetes.io/projected/47dc56bf-c795-49b1-ad78-c8407438f671-kube-api-access-sb248\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.636649 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-log-httpd\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.636750 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-config-data\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.738659 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.738705 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-run-httpd\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.738722 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.738754 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb248\" (UniqueName: \"kubernetes.io/projected/47dc56bf-c795-49b1-ad78-c8407438f671-kube-api-access-sb248\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.738799 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-log-httpd\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.738836 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-config-data\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.738917 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-scripts\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.739757 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-run-httpd\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.740566 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-log-httpd\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.749421 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-scripts\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.752903 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.755232 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.755685 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-config-data\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.785856 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb248\" (UniqueName: \"kubernetes.io/projected/47dc56bf-c795-49b1-ad78-c8407438f671-kube-api-access-sb248\") pod \"ceilometer-0\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " pod="openstack/ceilometer-0" Dec 04 14:13:27 crc kubenswrapper[4848]: I1204 14:13:27.952590 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.103707 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-65b648c688-s9c2l"] Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.114287 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-665fbf8d68-zvxk9"] Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.135366 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-79cb47cd96-sdhxb"] Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.136834 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.142627 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.145710 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.149513 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7fcddc797c-lm59t"] Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.151339 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.155218 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.155396 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.162629 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-79cb47cd96-sdhxb"] Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.200002 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7fcddc797c-lm59t"] Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.249346 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-combined-ca-bundle\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.249416 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-internal-tls-certs\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.249439 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-internal-tls-certs\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.249463 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-public-tls-certs\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.249488 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r6dh\" (UniqueName: \"kubernetes.io/projected/a1b263e8-986b-4822-9d9d-f1454757cadb-kube-api-access-9r6dh\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.249850 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-public-tls-certs\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.249969 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-combined-ca-bundle\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.250129 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data-custom\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.250765 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.250798 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data-custom\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.250871 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl6wc\" (UniqueName: \"kubernetes.io/projected/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-kube-api-access-gl6wc\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.250906 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353322 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-combined-ca-bundle\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353558 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-internal-tls-certs\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353583 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-internal-tls-certs\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353611 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-public-tls-certs\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353636 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r6dh\" (UniqueName: \"kubernetes.io/projected/a1b263e8-986b-4822-9d9d-f1454757cadb-kube-api-access-9r6dh\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353682 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-public-tls-certs\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353712 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-combined-ca-bundle\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353737 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data-custom\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353773 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353787 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data-custom\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353818 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl6wc\" (UniqueName: \"kubernetes.io/projected/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-kube-api-access-gl6wc\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.353837 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.358592 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.360854 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-public-tls-certs\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.361010 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-combined-ca-bundle\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.361389 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-combined-ca-bundle\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.361597 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-internal-tls-certs\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.362935 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-internal-tls-certs\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.364062 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data-custom\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.364352 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-public-tls-certs\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.364884 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data-custom\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.365176 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.376552 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl6wc\" (UniqueName: \"kubernetes.io/projected/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-kube-api-access-gl6wc\") pod \"heat-cfnapi-7fcddc797c-lm59t\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.378010 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r6dh\" (UniqueName: \"kubernetes.io/projected/a1b263e8-986b-4822-9d9d-f1454757cadb-kube-api-access-9r6dh\") pod \"heat-api-79cb47cd96-sdhxb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.409750 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a22994b-5397-4d0d-9d8a-0292eca01fb4" path="/var/lib/kubelet/pods/7a22994b-5397-4d0d-9d8a-0292eca01fb4/volumes" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.410451 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99e49830-7634-4257-af28-938f373af926" path="/var/lib/kubelet/pods/99e49830-7634-4257-af28-938f373af926/volumes" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.477207 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58569cf956-f89kq" event={"ID":"28dda188-1082-4b1b-a0dc-42297a0918b4","Type":"ContainerStarted","Data":"cb3e395fc1e3f68a5acfaa67d87dfabe72743efc0e5c614e2400a7e38283fa35"} Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.488606 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:28 crc kubenswrapper[4848]: I1204 14:13:28.512508 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.292317 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.188:9292/healthcheck\": read tcp 10.217.0.2:53742->10.217.0.188:9292: read: connection reset by peer" Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.292319 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.188:9292/healthcheck\": read tcp 10.217.0.2:53744->10.217.0.188:9292: read: connection reset by peer" Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.369216 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.369502 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" containerName="glance-log" containerID="cri-o://3befd5657f14087a56be9aa7c8ccf6b25b52ce5f18f74cc76d3b81e1e72396ae" gracePeriod=30 Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.369577 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" containerName="glance-httpd" containerID="cri-o://5c35ca14f75f66593173d79a1d0a76f32adfb5dbc514ced5be9ab4b76b525913" gracePeriod=30 Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.493080 4848 generic.go:334] "Generic (PLEG): container finished" podID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerID="099fcfe1d45f56ac6d66db57942f93cf89b334794714da0f9c2e4368156413a1" exitCode=0 Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.493160 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db70a1ef-eaa8-4da0-8abe-72aa5c822226","Type":"ContainerDied","Data":"099fcfe1d45f56ac6d66db57942f93cf89b334794714da0f9c2e4368156413a1"} Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.495150 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55dd7d4b8f-9bj88" event={"ID":"732df5f3-a0ba-449b-bf97-8604a569314f","Type":"ContainerStarted","Data":"ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1"} Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.495312 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:29 crc kubenswrapper[4848]: I1204 14:13:29.518754 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-55dd7d4b8f-9bj88" podStartSLOduration=4.5187352149999995 podStartE2EDuration="4.518735215s" podCreationTimestamp="2025-12-04 14:13:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:29.509128018 +0000 UTC m=+1513.451624546" watchObservedRunningTime="2025-12-04 14:13:29.518735215 +0000 UTC m=+1513.461231743" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.505589 4848 generic.go:334] "Generic (PLEG): container finished" podID="f252353f-4af7-445f-9860-30d4299d8ad6" containerID="3befd5657f14087a56be9aa7c8ccf6b25b52ce5f18f74cc76d3b81e1e72396ae" exitCode=143 Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.505702 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f252353f-4af7-445f-9860-30d4299d8ad6","Type":"ContainerDied","Data":"3befd5657f14087a56be9aa7c8ccf6b25b52ce5f18f74cc76d3b81e1e72396ae"} Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.717968 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.743941 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pk76z"] Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.745655 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.773161 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pk76z"] Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.825292 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlw88\" (UniqueName: \"kubernetes.io/projected/fcc6ad68-0f83-4a73-b83a-46eca178604e-kube-api-access-xlw88\") pod \"nova-api-db-create-pk76z\" (UID: \"fcc6ad68-0f83-4a73-b83a-46eca178604e\") " pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.825366 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fcc6ad68-0f83-4a73-b83a-46eca178604e-operator-scripts\") pod \"nova-api-db-create-pk76z\" (UID: \"fcc6ad68-0f83-4a73-b83a-46eca178604e\") " pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.841261 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3873-account-create-update-nc5n8"] Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.842823 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.846675 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.851937 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3873-account-create-update-nc5n8"] Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.860328 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-tn2wq"] Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.862649 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.904708 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tn2wq"] Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.937300 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thtvp\" (UniqueName: \"kubernetes.io/projected/34db090c-22b5-4462-a0ef-aca71a75a21a-kube-api-access-thtvp\") pod \"nova-api-3873-account-create-update-nc5n8\" (UID: \"34db090c-22b5-4462-a0ef-aca71a75a21a\") " pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.937544 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlw88\" (UniqueName: \"kubernetes.io/projected/fcc6ad68-0f83-4a73-b83a-46eca178604e-kube-api-access-xlw88\") pod \"nova-api-db-create-pk76z\" (UID: \"fcc6ad68-0f83-4a73-b83a-46eca178604e\") " pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.937666 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fcc6ad68-0f83-4a73-b83a-46eca178604e-operator-scripts\") pod \"nova-api-db-create-pk76z\" (UID: \"fcc6ad68-0f83-4a73-b83a-46eca178604e\") " pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.937863 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34db090c-22b5-4462-a0ef-aca71a75a21a-operator-scripts\") pod \"nova-api-3873-account-create-update-nc5n8\" (UID: \"34db090c-22b5-4462-a0ef-aca71a75a21a\") " pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.939303 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fcc6ad68-0f83-4a73-b83a-46eca178604e-operator-scripts\") pod \"nova-api-db-create-pk76z\" (UID: \"fcc6ad68-0f83-4a73-b83a-46eca178604e\") " pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:30 crc kubenswrapper[4848]: I1204 14:13:30.974004 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlw88\" (UniqueName: \"kubernetes.io/projected/fcc6ad68-0f83-4a73-b83a-46eca178604e-kube-api-access-xlw88\") pod \"nova-api-db-create-pk76z\" (UID: \"fcc6ad68-0f83-4a73-b83a-46eca178604e\") " pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.036140 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-6fc5j"] Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.037909 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.040398 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-operator-scripts\") pod \"nova-cell0-db-create-tn2wq\" (UID: \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\") " pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.040460 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thtvp\" (UniqueName: \"kubernetes.io/projected/34db090c-22b5-4462-a0ef-aca71a75a21a-kube-api-access-thtvp\") pod \"nova-api-3873-account-create-update-nc5n8\" (UID: \"34db090c-22b5-4462-a0ef-aca71a75a21a\") " pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.040483 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvlg6\" (UniqueName: \"kubernetes.io/projected/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-kube-api-access-zvlg6\") pod \"nova-cell0-db-create-tn2wq\" (UID: \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\") " pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.040644 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34db090c-22b5-4462-a0ef-aca71a75a21a-operator-scripts\") pod \"nova-api-3873-account-create-update-nc5n8\" (UID: \"34db090c-22b5-4462-a0ef-aca71a75a21a\") " pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.054314 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-e824-account-create-update-h466m"] Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.055776 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.058585 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34db090c-22b5-4462-a0ef-aca71a75a21a-operator-scripts\") pod \"nova-api-3873-account-create-update-nc5n8\" (UID: \"34db090c-22b5-4462-a0ef-aca71a75a21a\") " pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.060555 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.071780 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.077229 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e824-account-create-update-h466m"] Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.087832 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thtvp\" (UniqueName: \"kubernetes.io/projected/34db090c-22b5-4462-a0ef-aca71a75a21a-kube-api-access-thtvp\") pod \"nova-api-3873-account-create-update-nc5n8\" (UID: \"34db090c-22b5-4462-a0ef-aca71a75a21a\") " pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.092570 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6fc5j"] Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.143078 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnd8c\" (UniqueName: \"kubernetes.io/projected/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-kube-api-access-xnd8c\") pod \"nova-cell0-e824-account-create-update-h466m\" (UID: \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\") " pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.143177 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75e19b5f-83d0-4a02-8791-4da2749e0a12-operator-scripts\") pod \"nova-cell1-db-create-6fc5j\" (UID: \"75e19b5f-83d0-4a02-8791-4da2749e0a12\") " pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.143290 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-operator-scripts\") pod \"nova-cell0-e824-account-create-update-h466m\" (UID: \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\") " pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.143355 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lq2w\" (UniqueName: \"kubernetes.io/projected/75e19b5f-83d0-4a02-8791-4da2749e0a12-kube-api-access-8lq2w\") pod \"nova-cell1-db-create-6fc5j\" (UID: \"75e19b5f-83d0-4a02-8791-4da2749e0a12\") " pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.143387 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-operator-scripts\") pod \"nova-cell0-db-create-tn2wq\" (UID: \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\") " pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.143415 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvlg6\" (UniqueName: \"kubernetes.io/projected/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-kube-api-access-zvlg6\") pod \"nova-cell0-db-create-tn2wq\" (UID: \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\") " pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.149991 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.153832 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-operator-scripts\") pod \"nova-cell0-db-create-tn2wq\" (UID: \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\") " pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.166888 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvlg6\" (UniqueName: \"kubernetes.io/projected/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-kube-api-access-zvlg6\") pod \"nova-cell0-db-create-tn2wq\" (UID: \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\") " pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.177777 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-caff-account-create-update-sgpcw"] Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.179329 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.182003 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.185626 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.226154 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-caff-account-create-update-sgpcw"] Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.247614 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnd8c\" (UniqueName: \"kubernetes.io/projected/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-kube-api-access-xnd8c\") pod \"nova-cell0-e824-account-create-update-h466m\" (UID: \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\") " pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.247766 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75e19b5f-83d0-4a02-8791-4da2749e0a12-operator-scripts\") pod \"nova-cell1-db-create-6fc5j\" (UID: \"75e19b5f-83d0-4a02-8791-4da2749e0a12\") " pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.247848 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bd79f24-e81e-49bd-917d-5a6e3281f336-operator-scripts\") pod \"nova-cell1-caff-account-create-update-sgpcw\" (UID: \"2bd79f24-e81e-49bd-917d-5a6e3281f336\") " pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.247921 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-operator-scripts\") pod \"nova-cell0-e824-account-create-update-h466m\" (UID: \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\") " pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.248249 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lq2w\" (UniqueName: \"kubernetes.io/projected/75e19b5f-83d0-4a02-8791-4da2749e0a12-kube-api-access-8lq2w\") pod \"nova-cell1-db-create-6fc5j\" (UID: \"75e19b5f-83d0-4a02-8791-4da2749e0a12\") " pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.248314 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt75j\" (UniqueName: \"kubernetes.io/projected/2bd79f24-e81e-49bd-917d-5a6e3281f336-kube-api-access-wt75j\") pod \"nova-cell1-caff-account-create-update-sgpcw\" (UID: \"2bd79f24-e81e-49bd-917d-5a6e3281f336\") " pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.248689 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75e19b5f-83d0-4a02-8791-4da2749e0a12-operator-scripts\") pod \"nova-cell1-db-create-6fc5j\" (UID: \"75e19b5f-83d0-4a02-8791-4da2749e0a12\") " pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.249358 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-operator-scripts\") pod \"nova-cell0-e824-account-create-update-h466m\" (UID: \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\") " pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.284493 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lq2w\" (UniqueName: \"kubernetes.io/projected/75e19b5f-83d0-4a02-8791-4da2749e0a12-kube-api-access-8lq2w\") pod \"nova-cell1-db-create-6fc5j\" (UID: \"75e19b5f-83d0-4a02-8791-4da2749e0a12\") " pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.287118 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnd8c\" (UniqueName: \"kubernetes.io/projected/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-kube-api-access-xnd8c\") pod \"nova-cell0-e824-account-create-update-h466m\" (UID: \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\") " pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.350659 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bd79f24-e81e-49bd-917d-5a6e3281f336-operator-scripts\") pod \"nova-cell1-caff-account-create-update-sgpcw\" (UID: \"2bd79f24-e81e-49bd-917d-5a6e3281f336\") " pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.350860 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt75j\" (UniqueName: \"kubernetes.io/projected/2bd79f24-e81e-49bd-917d-5a6e3281f336-kube-api-access-wt75j\") pod \"nova-cell1-caff-account-create-update-sgpcw\" (UID: \"2bd79f24-e81e-49bd-917d-5a6e3281f336\") " pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.352158 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bd79f24-e81e-49bd-917d-5a6e3281f336-operator-scripts\") pod \"nova-cell1-caff-account-create-update-sgpcw\" (UID: \"2bd79f24-e81e-49bd-917d-5a6e3281f336\") " pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.387305 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt75j\" (UniqueName: \"kubernetes.io/projected/2bd79f24-e81e-49bd-917d-5a6e3281f336-kube-api-access-wt75j\") pod \"nova-cell1-caff-account-create-update-sgpcw\" (UID: \"2bd79f24-e81e-49bd-917d-5a6e3281f336\") " pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.510560 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.535704 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.549050 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.554468 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.676929 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-combined-ca-bundle\") pod \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.676989 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cl5d\" (UniqueName: \"kubernetes.io/projected/db70a1ef-eaa8-4da0-8abe-72aa5c822226-kube-api-access-2cl5d\") pod \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.677031 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-httpd-run\") pod \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.677073 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-logs\") pod \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.677165 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.677295 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-config-data\") pod \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.677426 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-scripts\") pod \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.677581 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-public-tls-certs\") pod \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\" (UID: \"db70a1ef-eaa8-4da0-8abe-72aa5c822226\") " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.679061 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-logs" (OuterVolumeSpecName: "logs") pod "db70a1ef-eaa8-4da0-8abe-72aa5c822226" (UID: "db70a1ef-eaa8-4da0-8abe-72aa5c822226"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.679255 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "db70a1ef-eaa8-4da0-8abe-72aa5c822226" (UID: "db70a1ef-eaa8-4da0-8abe-72aa5c822226"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.692393 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-scripts" (OuterVolumeSpecName: "scripts") pod "db70a1ef-eaa8-4da0-8abe-72aa5c822226" (UID: "db70a1ef-eaa8-4da0-8abe-72aa5c822226"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.692536 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db70a1ef-eaa8-4da0-8abe-72aa5c822226-kube-api-access-2cl5d" (OuterVolumeSpecName: "kube-api-access-2cl5d") pod "db70a1ef-eaa8-4da0-8abe-72aa5c822226" (UID: "db70a1ef-eaa8-4da0-8abe-72aa5c822226"). InnerVolumeSpecName "kube-api-access-2cl5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.694219 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "db70a1ef-eaa8-4da0-8abe-72aa5c822226" (UID: "db70a1ef-eaa8-4da0-8abe-72aa5c822226"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.742090 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db70a1ef-eaa8-4da0-8abe-72aa5c822226" (UID: "db70a1ef-eaa8-4da0-8abe-72aa5c822226"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.781587 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.781616 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cl5d\" (UniqueName: \"kubernetes.io/projected/db70a1ef-eaa8-4da0-8abe-72aa5c822226-kube-api-access-2cl5d\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.781627 4848 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.781636 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db70a1ef-eaa8-4da0-8abe-72aa5c822226-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.781658 4848 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.781666 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.798706 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "db70a1ef-eaa8-4da0-8abe-72aa5c822226" (UID: "db70a1ef-eaa8-4da0-8abe-72aa5c822226"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.814091 4848 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.832079 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-config-data" (OuterVolumeSpecName: "config-data") pod "db70a1ef-eaa8-4da0-8abe-72aa5c822226" (UID: "db70a1ef-eaa8-4da0-8abe-72aa5c822226"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.832129 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7fcddc797c-lm59t"] Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.884831 4848 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.884872 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.884886 4848 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db70a1ef-eaa8-4da0-8abe-72aa5c822226-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:31 crc kubenswrapper[4848]: I1204 14:13:31.970847 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.004943 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-79cb47cd96-sdhxb"] Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.194353 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.198601 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-b65c5f4cf-q22f8" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.563249 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pk76z"] Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.572108 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-65b648c688-s9c2l" event={"ID":"70bfcde6-2ad2-44fc-ae39-5821c63f4918","Type":"ContainerStarted","Data":"4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc"} Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.572268 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-65b648c688-s9c2l" podUID="70bfcde6-2ad2-44fc-ae39-5821c63f4918" containerName="heat-api" containerID="cri-o://4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc" gracePeriod=60 Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.572539 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.581838 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" event={"ID":"b38b0b7c-d339-418d-b2c4-0278a451626a","Type":"ContainerStarted","Data":"27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa"} Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.582018 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" podUID="b38b0b7c-d339-418d-b2c4-0278a451626a" containerName="heat-cfnapi" containerID="cri-o://27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa" gracePeriod=60 Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.582118 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.591071 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6fc5j"] Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.596330 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerStarted","Data":"bd2741cbb4367c200745e3e5a145842265cf81aa8b51d8c6bef6cb392c5e6873"} Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.601683 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3873-account-create-update-nc5n8"] Dec 04 14:13:32 crc kubenswrapper[4848]: W1204 14:13:32.606623 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75e19b5f_83d0_4a02_8791_4da2749e0a12.slice/crio-3b41e6d4d8cffe93315d37215100bd1db5ff27556ce1f8d5cf1fd5428183b375 WatchSource:0}: Error finding container 3b41e6d4d8cffe93315d37215100bd1db5ff27556ce1f8d5cf1fd5428183b375: Status 404 returned error can't find the container with id 3b41e6d4d8cffe93315d37215100bd1db5ff27556ce1f8d5cf1fd5428183b375 Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.610421 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-65b648c688-s9c2l" podStartSLOduration=8.035428404 podStartE2EDuration="13.610403387s" podCreationTimestamp="2025-12-04 14:13:19 +0000 UTC" firstStartedPulling="2025-12-04 14:13:25.626372215 +0000 UTC m=+1509.568868743" lastFinishedPulling="2025-12-04 14:13:31.201347198 +0000 UTC m=+1515.143843726" observedRunningTime="2025-12-04 14:13:32.591464408 +0000 UTC m=+1516.533960936" watchObservedRunningTime="2025-12-04 14:13:32.610403387 +0000 UTC m=+1516.552899915" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.612137 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tn2wq"] Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.620322 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.620476 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"db70a1ef-eaa8-4da0-8abe-72aa5c822226","Type":"ContainerDied","Data":"bc6faf4dc722e4aaedc09369cabab98084d15e87c650dbb55ad5712971365e06"} Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.620535 4848 scope.go:117] "RemoveContainer" containerID="099fcfe1d45f56ac6d66db57942f93cf89b334794714da0f9c2e4368156413a1" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.625736 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" event={"ID":"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9","Type":"ContainerStarted","Data":"b0b68ef983defd475a7df4577f43b9ef21339fafa35e48d88b54116bd9824fe3"} Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.628880 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-79cb47cd96-sdhxb" event={"ID":"a1b263e8-986b-4822-9d9d-f1454757cadb","Type":"ContainerStarted","Data":"876713e322f8762625d5bf3b53d1f7702866228ec46495334ae74632a3dc3feb"} Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.652439 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" podStartSLOduration=8.092780351 podStartE2EDuration="13.652423197s" podCreationTimestamp="2025-12-04 14:13:19 +0000 UTC" firstStartedPulling="2025-12-04 14:13:25.684647042 +0000 UTC m=+1509.627143570" lastFinishedPulling="2025-12-04 14:13:31.244289888 +0000 UTC m=+1515.186786416" observedRunningTime="2025-12-04 14:13:32.620042125 +0000 UTC m=+1516.562538663" watchObservedRunningTime="2025-12-04 14:13:32.652423197 +0000 UTC m=+1516.594919725" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.721059 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.739004 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.806017 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:13:32 crc kubenswrapper[4848]: E1204 14:13:32.807166 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-log" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.807181 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-log" Dec 04 14:13:32 crc kubenswrapper[4848]: E1204 14:13:32.807226 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-httpd" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.807232 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-httpd" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.807718 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-log" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.807770 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" containerName="glance-httpd" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.854502 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.860514 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.860751 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.881320 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.925632 4848 scope.go:117] "RemoveContainer" containerID="c7727796ae2e569ffb4887e1d8acaaae87939cc44b10b82cc3a84810cd2191aa" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.969667 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.969898 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.970036 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.970175 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.970203 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.970350 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd8zc\" (UniqueName: \"kubernetes.io/projected/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-kube-api-access-zd8zc\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.970384 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-logs\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:32 crc kubenswrapper[4848]: I1204 14:13:32.970482 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.073313 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd8zc\" (UniqueName: \"kubernetes.io/projected/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-kube-api-access-zd8zc\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.073354 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-logs\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.073407 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.073463 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.073485 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.073526 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.073578 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.073595 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.074809 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.080909 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-logs\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.083775 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.093188 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.099030 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.102361 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.105360 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd8zc\" (UniqueName: \"kubernetes.io/projected/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-kube-api-access-zd8zc\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.109660 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5008baa-3c8b-4048-a87e-4e6f3e5d2327-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.161450 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e824-account-create-update-h466m"] Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.180507 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d5008baa-3c8b-4048-a87e-4e6f3e5d2327\") " pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.201134 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-caff-account-create-update-sgpcw"] Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.433227 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.665386 4848 generic.go:334] "Generic (PLEG): container finished" podID="28dda188-1082-4b1b-a0dc-42297a0918b4" containerID="a5c7b2c05c03f602c56b8fa91b462d1560229afd2a263d198ca9912ed1c90df8" exitCode=1 Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.665811 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58569cf956-f89kq" event={"ID":"28dda188-1082-4b1b-a0dc-42297a0918b4","Type":"ContainerDied","Data":"a5c7b2c05c03f602c56b8fa91b462d1560229afd2a263d198ca9912ed1c90df8"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.666631 4848 scope.go:117] "RemoveContainer" containerID="a5c7b2c05c03f602c56b8fa91b462d1560229afd2a263d198ca9912ed1c90df8" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.707545 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" event={"ID":"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9","Type":"ContainerStarted","Data":"a74545ca7904da03f0d3b9228bbc0239cbf44b77771af84bf1a6800ed27f0255"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.708120 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.729896 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.730021 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6fc5j" event={"ID":"75e19b5f-83d0-4a02-8791-4da2749e0a12","Type":"ContainerStarted","Data":"3b41e6d4d8cffe93315d37215100bd1db5ff27556ce1f8d5cf1fd5428183b375"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.742663 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" podStartSLOduration=5.74264708 podStartE2EDuration="5.74264708s" podCreationTimestamp="2025-12-04 14:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:33.74177863 +0000 UTC m=+1517.684275168" watchObservedRunningTime="2025-12-04 14:13:33.74264708 +0000 UTC m=+1517.685143608" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.749237 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e824-account-create-update-h466m" event={"ID":"d94a8c9d-5f73-44fa-8fc7-f725f21a8127","Type":"ContainerStarted","Data":"93ae6e3ee3c4da480f880422b4ba0ee42dfac72536f71893de57fc053da54901"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.787787 4848 generic.go:334] "Generic (PLEG): container finished" podID="f252353f-4af7-445f-9860-30d4299d8ad6" containerID="5c35ca14f75f66593173d79a1d0a76f32adfb5dbc514ced5be9ab4b76b525913" exitCode=0 Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.787871 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f252353f-4af7-445f-9860-30d4299d8ad6","Type":"ContainerDied","Data":"5c35ca14f75f66593173d79a1d0a76f32adfb5dbc514ced5be9ab4b76b525913"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.787905 4848 scope.go:117] "RemoveContainer" containerID="5c35ca14f75f66593173d79a1d0a76f32adfb5dbc514ced5be9ab4b76b525913" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.788045 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.827759 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-caff-account-create-update-sgpcw" event={"ID":"2bd79f24-e81e-49bd-917d-5a6e3281f336","Type":"ContainerStarted","Data":"1ae3300fa5c2605e28ea6fe281136d4ad37dd1b115ffb6f578de0f55fab4a804"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.837352 4848 generic.go:334] "Generic (PLEG): container finished" podID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" containerID="5f9a0273ec472ab5dcee58b582ad85814381177accb804c85667c6157b21effa" exitCode=1 Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.837617 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-878c6546-2c98w" event={"ID":"a9478f9e-a08b-4d95-b57a-fb8ed97850fb","Type":"ContainerDied","Data":"5f9a0273ec472ab5dcee58b582ad85814381177accb804c85667c6157b21effa"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.838216 4848 scope.go:117] "RemoveContainer" containerID="5f9a0273ec472ab5dcee58b582ad85814381177accb804c85667c6157b21effa" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.848419 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-79cb47cd96-sdhxb" event={"ID":"a1b263e8-986b-4822-9d9d-f1454757cadb","Type":"ContainerStarted","Data":"1ee1caa64261f0dbf0c1ccd8ba30c50baaf85f0078f6881ee025818185371175"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.849621 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.872463 4848 scope.go:117] "RemoveContainer" containerID="3befd5657f14087a56be9aa7c8ccf6b25b52ce5f18f74cc76d3b81e1e72396ae" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.885337 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3873-account-create-update-nc5n8" event={"ID":"34db090c-22b5-4462-a0ef-aca71a75a21a","Type":"ContainerStarted","Data":"ea2f66ea4dd001e26b15e3525988cf1b5ea400a47b0835c20da0ebdb97143c72"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.907609 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tn2wq" event={"ID":"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232","Type":"ContainerStarted","Data":"c21c9674ef4096b737baf72b4a540f8f36a04aa83bf95816d75f616b50a0a26c"} Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.928675 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-internal-tls-certs\") pod \"f252353f-4af7-445f-9860-30d4299d8ad6\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.928809 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-config-data\") pod \"f252353f-4af7-445f-9860-30d4299d8ad6\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.928972 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-combined-ca-bundle\") pod \"f252353f-4af7-445f-9860-30d4299d8ad6\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.929065 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f252353f-4af7-445f-9860-30d4299d8ad6\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.929115 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-scripts\") pod \"f252353f-4af7-445f-9860-30d4299d8ad6\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.929141 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-logs\") pod \"f252353f-4af7-445f-9860-30d4299d8ad6\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.929165 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2gcv\" (UniqueName: \"kubernetes.io/projected/f252353f-4af7-445f-9860-30d4299d8ad6-kube-api-access-j2gcv\") pod \"f252353f-4af7-445f-9860-30d4299d8ad6\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.929196 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-httpd-run\") pod \"f252353f-4af7-445f-9860-30d4299d8ad6\" (UID: \"f252353f-4af7-445f-9860-30d4299d8ad6\") " Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.930459 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-logs" (OuterVolumeSpecName: "logs") pod "f252353f-4af7-445f-9860-30d4299d8ad6" (UID: "f252353f-4af7-445f-9860-30d4299d8ad6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.937856 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "f252353f-4af7-445f-9860-30d4299d8ad6" (UID: "f252353f-4af7-445f-9860-30d4299d8ad6"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.938181 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f252353f-4af7-445f-9860-30d4299d8ad6" (UID: "f252353f-4af7-445f-9860-30d4299d8ad6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.943857 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f252353f-4af7-445f-9860-30d4299d8ad6-kube-api-access-j2gcv" (OuterVolumeSpecName: "kube-api-access-j2gcv") pod "f252353f-4af7-445f-9860-30d4299d8ad6" (UID: "f252353f-4af7-445f-9860-30d4299d8ad6"). InnerVolumeSpecName "kube-api-access-j2gcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.946336 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-scripts" (OuterVolumeSpecName: "scripts") pod "f252353f-4af7-445f-9860-30d4299d8ad6" (UID: "f252353f-4af7-445f-9860-30d4299d8ad6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.981692 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-79cb47cd96-sdhxb" podStartSLOduration=5.981492028 podStartE2EDuration="5.981492028s" podCreationTimestamp="2025-12-04 14:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:33.918699478 +0000 UTC m=+1517.861196006" watchObservedRunningTime="2025-12-04 14:13:33.981492028 +0000 UTC m=+1517.923988546" Dec 04 14:13:33 crc kubenswrapper[4848]: I1204 14:13:33.983345 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pk76z" event={"ID":"fcc6ad68-0f83-4a73-b83a-46eca178604e","Type":"ContainerStarted","Data":"107b49ae018da26dbe839a46e23c34ee287b9b6f89fc88957b3d710bd2071b45"} Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.000107 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f252353f-4af7-445f-9860-30d4299d8ad6" (UID: "f252353f-4af7-445f-9860-30d4299d8ad6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.057468 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.057503 4848 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.058411 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.058427 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.058439 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2gcv\" (UniqueName: \"kubernetes.io/projected/f252353f-4af7-445f-9860-30d4299d8ad6-kube-api-access-j2gcv\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.058447 4848 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f252353f-4af7-445f-9860-30d4299d8ad6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.072631 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-3873-account-create-update-nc5n8" podStartSLOduration=4.072613027 podStartE2EDuration="4.072613027s" podCreationTimestamp="2025-12-04 14:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:33.944381549 +0000 UTC m=+1517.886878087" watchObservedRunningTime="2025-12-04 14:13:34.072613027 +0000 UTC m=+1518.015109555" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.088503 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-config-data" (OuterVolumeSpecName: "config-data") pod "f252353f-4af7-445f-9860-30d4299d8ad6" (UID: "f252353f-4af7-445f-9860-30d4299d8ad6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.100751 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-pk76z" podStartSLOduration=4.100730473 podStartE2EDuration="4.100730473s" podCreationTimestamp="2025-12-04 14:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:33.99615311 +0000 UTC m=+1517.938649638" watchObservedRunningTime="2025-12-04 14:13:34.100730473 +0000 UTC m=+1518.043227001" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.137914 4848 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.166420 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.166446 4848 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.182769 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vlctk"] Dec 04 14:13:34 crc kubenswrapper[4848]: E1204 14:13:34.184152 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" containerName="glance-httpd" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.184169 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" containerName="glance-httpd" Dec 04 14:13:34 crc kubenswrapper[4848]: E1204 14:13:34.184186 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" containerName="glance-log" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.184192 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" containerName="glance-log" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.184438 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" containerName="glance-httpd" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.184454 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" containerName="glance-log" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.187672 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.208959 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vlctk"] Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.342428 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f252353f-4af7-445f-9860-30d4299d8ad6" (UID: "f252353f-4af7-445f-9860-30d4299d8ad6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.348021 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.374503 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkbxw\" (UniqueName: \"kubernetes.io/projected/c0f69f89-4463-4f8a-b892-ca81630f4b67-kube-api-access-kkbxw\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.374706 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-catalog-content\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.374729 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-utilities\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.374787 4848 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f252353f-4af7-445f-9860-30d4299d8ad6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.435276 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db70a1ef-eaa8-4da0-8abe-72aa5c822226" path="/var/lib/kubelet/pods/db70a1ef-eaa8-4da0-8abe-72aa5c822226/volumes" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.477206 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-catalog-content\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.477241 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-utilities\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.477297 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkbxw\" (UniqueName: \"kubernetes.io/projected/c0f69f89-4463-4f8a-b892-ca81630f4b67-kube-api-access-kkbxw\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.477711 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-catalog-content\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.477771 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-utilities\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.495640 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkbxw\" (UniqueName: \"kubernetes.io/projected/c0f69f89-4463-4f8a-b892-ca81630f4b67-kube-api-access-kkbxw\") pod \"certified-operators-vlctk\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.795980 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.841147 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:13:34 crc kubenswrapper[4848]: E1204 14:13:34.867491 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bd79f24_e81e_49bd_917d_5a6e3281f336.slice/crio-conmon-c940759c43b01c94d60d81a1069f82c0b54af5b13c2891edb642549a6774735a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bd79f24_e81e_49bd_917d_5a6e3281f336.slice/crio-c940759c43b01c94d60d81a1069f82c0b54af5b13c2891edb642549a6774735a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd94a8c9d_5f73_44fa_8fc7_f725f21a8127.slice/crio-7ab4def029874b47a7238ffe5351fd78acdc1d4792b252fd28ff5214af48c1ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd94a8c9d_5f73_44fa_8fc7_f725f21a8127.slice/crio-conmon-7ab4def029874b47a7238ffe5351fd78acdc1d4792b252fd28ff5214af48c1ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34db090c_22b5_4462_a0ef_aca71a75a21a.slice/crio-conmon-bc9161643dba2afa8240d7ac32bb8827495bd79e26bd160d7ba4c91469643dc2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34db090c_22b5_4462_a0ef_aca71a75a21a.slice/crio-bc9161643dba2afa8240d7ac32bb8827495bd79e26bd160d7ba4c91469643dc2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75e19b5f_83d0_4a02_8791_4da2749e0a12.slice/crio-e8eee4a623c4f7cb747ee23a1e57a2c62d6eb3b9701ec420784603cdcbac0d12.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75e19b5f_83d0_4a02_8791_4da2749e0a12.slice/crio-conmon-e8eee4a623c4f7cb747ee23a1e57a2c62d6eb3b9701ec420784603cdcbac0d12.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf692bc07_eca1_4b0b_b5fc_a9ad3aa34232.slice/crio-conmon-272ebb608ee1f2bb5c58907aab4f4947a04ae59529614952c2994ff490528f1b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcc6ad68_0f83_4a73_b83a_46eca178604e.slice/crio-10996e0c1d719def5deae568a099963cc051c51e42b330c75f0b28c900a1333d.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.901344 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.918719 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.921371 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.924940 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.925705 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 14:13:34 crc kubenswrapper[4848]: I1204 14:13:34.927738 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.025496 4848 generic.go:334] "Generic (PLEG): container finished" podID="34db090c-22b5-4462-a0ef-aca71a75a21a" containerID="bc9161643dba2afa8240d7ac32bb8827495bd79e26bd160d7ba4c91469643dc2" exitCode=0 Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.025565 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3873-account-create-update-nc5n8" event={"ID":"34db090c-22b5-4462-a0ef-aca71a75a21a","Type":"ContainerDied","Data":"bc9161643dba2afa8240d7ac32bb8827495bd79e26bd160d7ba4c91469643dc2"} Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.037677 4848 generic.go:334] "Generic (PLEG): container finished" podID="75e19b5f-83d0-4a02-8791-4da2749e0a12" containerID="e8eee4a623c4f7cb747ee23a1e57a2c62d6eb3b9701ec420784603cdcbac0d12" exitCode=0 Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.037756 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6fc5j" event={"ID":"75e19b5f-83d0-4a02-8791-4da2749e0a12","Type":"ContainerDied","Data":"e8eee4a623c4f7cb747ee23a1e57a2c62d6eb3b9701ec420784603cdcbac0d12"} Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.079660 4848 generic.go:334] "Generic (PLEG): container finished" podID="d94a8c9d-5f73-44fa-8fc7-f725f21a8127" containerID="7ab4def029874b47a7238ffe5351fd78acdc1d4792b252fd28ff5214af48c1ca" exitCode=0 Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.079741 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e824-account-create-update-h466m" event={"ID":"d94a8c9d-5f73-44fa-8fc7-f725f21a8127","Type":"ContainerDied","Data":"7ab4def029874b47a7238ffe5351fd78acdc1d4792b252fd28ff5214af48c1ca"} Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.094155 4848 generic.go:334] "Generic (PLEG): container finished" podID="fcc6ad68-0f83-4a73-b83a-46eca178604e" containerID="10996e0c1d719def5deae568a099963cc051c51e42b330c75f0b28c900a1333d" exitCode=0 Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.094213 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pk76z" event={"ID":"fcc6ad68-0f83-4a73-b83a-46eca178604e","Type":"ContainerDied","Data":"10996e0c1d719def5deae568a099963cc051c51e42b330c75f0b28c900a1333d"} Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.101293 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5008baa-3c8b-4048-a87e-4e6f3e5d2327","Type":"ContainerStarted","Data":"58e1c354f103f3dade7970e237d425135adea02d96a6f11dcc01d37a7e39de72"} Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.123685 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.123747 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.123777 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c33bbab-5535-4d34-8b87-90c7e1549281-logs\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.123816 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np6gf\" (UniqueName: \"kubernetes.io/projected/0c33bbab-5535-4d34-8b87-90c7e1549281-kube-api-access-np6gf\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.123872 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.123908 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.123922 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c33bbab-5535-4d34-8b87-90c7e1549281-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.123977 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.130100 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerStarted","Data":"564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6"} Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.141224 4848 generic.go:334] "Generic (PLEG): container finished" podID="2bd79f24-e81e-49bd-917d-5a6e3281f336" containerID="c940759c43b01c94d60d81a1069f82c0b54af5b13c2891edb642549a6774735a" exitCode=0 Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.141302 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-caff-account-create-update-sgpcw" event={"ID":"2bd79f24-e81e-49bd-917d-5a6e3281f336","Type":"ContainerDied","Data":"c940759c43b01c94d60d81a1069f82c0b54af5b13c2891edb642549a6774735a"} Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.185434 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.188655 4848 generic.go:334] "Generic (PLEG): container finished" podID="f692bc07-eca1-4b0b-b5fc-a9ad3aa34232" containerID="272ebb608ee1f2bb5c58907aab4f4947a04ae59529614952c2994ff490528f1b" exitCode=0 Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.188714 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tn2wq" event={"ID":"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232","Type":"ContainerDied","Data":"272ebb608ee1f2bb5c58907aab4f4947a04ae59529614952c2994ff490528f1b"} Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.225938 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.226013 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.226038 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c33bbab-5535-4d34-8b87-90c7e1549281-logs\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.226080 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np6gf\" (UniqueName: \"kubernetes.io/projected/0c33bbab-5535-4d34-8b87-90c7e1549281-kube-api-access-np6gf\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.226135 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.226171 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.226186 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c33bbab-5535-4d34-8b87-90c7e1549281-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.226218 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.228519 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.239509 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c33bbab-5535-4d34-8b87-90c7e1549281-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.255423 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.256677 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c33bbab-5535-4d34-8b87-90c7e1549281-logs\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.277655 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.278099 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.286014 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np6gf\" (UniqueName: \"kubernetes.io/projected/0c33bbab-5535-4d34-8b87-90c7e1549281-kube-api-access-np6gf\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.315484 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kbh4n"] Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.315557 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c33bbab-5535-4d34-8b87-90c7e1549281-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.315714 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" podUID="88b349db-6f99-4351-8a39-7dc1ba31496d" containerName="dnsmasq-dns" containerID="cri-o://498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e" gracePeriod=10 Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.373029 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"0c33bbab-5535-4d34-8b87-90c7e1549281\") " pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.616682 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:35 crc kubenswrapper[4848]: I1204 14:13:35.739175 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vlctk"] Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.242633 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerStarted","Data":"d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c"} Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.243541 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.300548 4848 generic.go:334] "Generic (PLEG): container finished" podID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" containerID="c4f0e98194b7e03b78802853c99d8014a587b5764feac053e0b254bdcc6c066a" exitCode=1 Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.300837 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-878c6546-2c98w" event={"ID":"a9478f9e-a08b-4d95-b57a-fb8ed97850fb","Type":"ContainerDied","Data":"c4f0e98194b7e03b78802853c99d8014a587b5764feac053e0b254bdcc6c066a"} Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.300911 4848 scope.go:117] "RemoveContainer" containerID="5f9a0273ec472ab5dcee58b582ad85814381177accb804c85667c6157b21effa" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.302860 4848 scope.go:117] "RemoveContainer" containerID="c4f0e98194b7e03b78802853c99d8014a587b5764feac053e0b254bdcc6c066a" Dec 04 14:13:36 crc kubenswrapper[4848]: E1204 14:13:36.303668 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-878c6546-2c98w_openstack(a9478f9e-a08b-4d95-b57a-fb8ed97850fb)\"" pod="openstack/heat-api-878c6546-2c98w" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.355301 4848 generic.go:334] "Generic (PLEG): container finished" podID="28dda188-1082-4b1b-a0dc-42297a0918b4" containerID="85e1878ee8070715c8ff9e255a06174863bb49f388e2ebceb783c5bef7f70b5b" exitCode=1 Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.355385 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58569cf956-f89kq" event={"ID":"28dda188-1082-4b1b-a0dc-42297a0918b4","Type":"ContainerDied","Data":"85e1878ee8070715c8ff9e255a06174863bb49f388e2ebceb783c5bef7f70b5b"} Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.376365 4848 scope.go:117] "RemoveContainer" containerID="85e1878ee8070715c8ff9e255a06174863bb49f388e2ebceb783c5bef7f70b5b" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.377247 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-svc\") pod \"88b349db-6f99-4351-8a39-7dc1ba31496d\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.377342 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-swift-storage-0\") pod \"88b349db-6f99-4351-8a39-7dc1ba31496d\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.377420 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-sb\") pod \"88b349db-6f99-4351-8a39-7dc1ba31496d\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.377531 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-config\") pod \"88b349db-6f99-4351-8a39-7dc1ba31496d\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " Dec 04 14:13:36 crc kubenswrapper[4848]: E1204 14:13:36.377836 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58569cf956-f89kq_openstack(28dda188-1082-4b1b-a0dc-42297a0918b4)\"" pod="openstack/heat-cfnapi-58569cf956-f89kq" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.378092 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwx8r\" (UniqueName: \"kubernetes.io/projected/88b349db-6f99-4351-8a39-7dc1ba31496d-kube-api-access-hwx8r\") pod \"88b349db-6f99-4351-8a39-7dc1ba31496d\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.378182 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-nb\") pod \"88b349db-6f99-4351-8a39-7dc1ba31496d\" (UID: \"88b349db-6f99-4351-8a39-7dc1ba31496d\") " Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.395773 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b349db-6f99-4351-8a39-7dc1ba31496d-kube-api-access-hwx8r" (OuterVolumeSpecName: "kube-api-access-hwx8r") pod "88b349db-6f99-4351-8a39-7dc1ba31496d" (UID: "88b349db-6f99-4351-8a39-7dc1ba31496d"). InnerVolumeSpecName "kube-api-access-hwx8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.452069 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f252353f-4af7-445f-9860-30d4299d8ad6" path="/var/lib/kubelet/pods/f252353f-4af7-445f-9860-30d4299d8ad6/volumes" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.452387 4848 generic.go:334] "Generic (PLEG): container finished" podID="88b349db-6f99-4351-8a39-7dc1ba31496d" containerID="498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e" exitCode=0 Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.452537 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.460746 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vlctk" event={"ID":"c0f69f89-4463-4f8a-b892-ca81630f4b67","Type":"ContainerStarted","Data":"c30c9472e3d5249c219f4d2aebf369e794d2725804a0d1d95e4bcea870923fd0"} Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.460783 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" event={"ID":"88b349db-6f99-4351-8a39-7dc1ba31496d","Type":"ContainerDied","Data":"498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e"} Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.460799 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kbh4n" event={"ID":"88b349db-6f99-4351-8a39-7dc1ba31496d","Type":"ContainerDied","Data":"3b5901029957843d3bd29a36cf59301d0d1cf99518d52242cf4731d06e30aa5f"} Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.463229 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.463686 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5008baa-3c8b-4048-a87e-4e6f3e5d2327","Type":"ContainerStarted","Data":"f445d194e4c80abfcefca939b5ed952be9d6a4292ce85254a9c9d602619592a4"} Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.481833 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwx8r\" (UniqueName: \"kubernetes.io/projected/88b349db-6f99-4351-8a39-7dc1ba31496d-kube-api-access-hwx8r\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.515474 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "88b349db-6f99-4351-8a39-7dc1ba31496d" (UID: "88b349db-6f99-4351-8a39-7dc1ba31496d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.535834 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "88b349db-6f99-4351-8a39-7dc1ba31496d" (UID: "88b349db-6f99-4351-8a39-7dc1ba31496d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.548198 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "88b349db-6f99-4351-8a39-7dc1ba31496d" (UID: "88b349db-6f99-4351-8a39-7dc1ba31496d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.549180 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-config" (OuterVolumeSpecName: "config") pod "88b349db-6f99-4351-8a39-7dc1ba31496d" (UID: "88b349db-6f99-4351-8a39-7dc1ba31496d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.554905 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "88b349db-6f99-4351-8a39-7dc1ba31496d" (UID: "88b349db-6f99-4351-8a39-7dc1ba31496d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.576829 4848 scope.go:117] "RemoveContainer" containerID="a5c7b2c05c03f602c56b8fa91b462d1560229afd2a263d198ca9912ed1c90df8" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.585681 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.585719 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.585736 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.585748 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.585760 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88b349db-6f99-4351-8a39-7dc1ba31496d-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.627115 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.627180 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.652774 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.653319 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.781103 4848 scope.go:117] "RemoveContainer" containerID="498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.797247 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kbh4n"] Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.814903 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kbh4n"] Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.856255 4848 scope.go:117] "RemoveContainer" containerID="6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.974244 4848 scope.go:117] "RemoveContainer" containerID="498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e" Dec 04 14:13:36 crc kubenswrapper[4848]: E1204 14:13:36.978061 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e\": container with ID starting with 498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e not found: ID does not exist" containerID="498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.978126 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e"} err="failed to get container status \"498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e\": rpc error: code = NotFound desc = could not find container \"498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e\": container with ID starting with 498ca0e4cc0585fe35d8f6ae007ceeb099873b5e7402e15e9bdcfe88a34e875e not found: ID does not exist" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.978150 4848 scope.go:117] "RemoveContainer" containerID="6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f" Dec 04 14:13:36 crc kubenswrapper[4848]: E1204 14:13:36.986539 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f\": container with ID starting with 6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f not found: ID does not exist" containerID="6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f" Dec 04 14:13:36 crc kubenswrapper[4848]: I1204 14:13:36.986579 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f"} err="failed to get container status \"6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f\": rpc error: code = NotFound desc = could not find container \"6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f\": container with ID starting with 6440fb4747c3fa3d28c19b2b30456bcc16f944525846e6951cf60dff9a8d375f not found: ID does not exist" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.096818 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.203689 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvlg6\" (UniqueName: \"kubernetes.io/projected/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-kube-api-access-zvlg6\") pod \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\" (UID: \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.203764 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-operator-scripts\") pod \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\" (UID: \"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.204603 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f692bc07-eca1-4b0b-b5fc-a9ad3aa34232" (UID: "f692bc07-eca1-4b0b-b5fc-a9ad3aa34232"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.232197 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-kube-api-access-zvlg6" (OuterVolumeSpecName: "kube-api-access-zvlg6") pod "f692bc07-eca1-4b0b-b5fc-a9ad3aa34232" (UID: "f692bc07-eca1-4b0b-b5fc-a9ad3aa34232"). InnerVolumeSpecName "kube-api-access-zvlg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.311422 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvlg6\" (UniqueName: \"kubernetes.io/projected/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-kube-api-access-zvlg6\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.311446 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.322382 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.394648 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.419127 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-operator-scripts\") pod \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\" (UID: \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.419298 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnd8c\" (UniqueName: \"kubernetes.io/projected/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-kube-api-access-xnd8c\") pod \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\" (UID: \"d94a8c9d-5f73-44fa-8fc7-f725f21a8127\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.420122 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d94a8c9d-5f73-44fa-8fc7-f725f21a8127" (UID: "d94a8c9d-5f73-44fa-8fc7-f725f21a8127"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.420377 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.423493 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-kube-api-access-xnd8c" (OuterVolumeSpecName: "kube-api-access-xnd8c") pod "d94a8c9d-5f73-44fa-8fc7-f725f21a8127" (UID: "d94a8c9d-5f73-44fa-8fc7-f725f21a8127"). InnerVolumeSpecName "kube-api-access-xnd8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.427360 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.455172 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.470436 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.523159 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75e19b5f-83d0-4a02-8791-4da2749e0a12-operator-scripts\") pod \"75e19b5f-83d0-4a02-8791-4da2749e0a12\" (UID: \"75e19b5f-83d0-4a02-8791-4da2749e0a12\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.523236 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34db090c-22b5-4462-a0ef-aca71a75a21a-operator-scripts\") pod \"34db090c-22b5-4462-a0ef-aca71a75a21a\" (UID: \"34db090c-22b5-4462-a0ef-aca71a75a21a\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.523292 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt75j\" (UniqueName: \"kubernetes.io/projected/2bd79f24-e81e-49bd-917d-5a6e3281f336-kube-api-access-wt75j\") pod \"2bd79f24-e81e-49bd-917d-5a6e3281f336\" (UID: \"2bd79f24-e81e-49bd-917d-5a6e3281f336\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.523376 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thtvp\" (UniqueName: \"kubernetes.io/projected/34db090c-22b5-4462-a0ef-aca71a75a21a-kube-api-access-thtvp\") pod \"34db090c-22b5-4462-a0ef-aca71a75a21a\" (UID: \"34db090c-22b5-4462-a0ef-aca71a75a21a\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.523400 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bd79f24-e81e-49bd-917d-5a6e3281f336-operator-scripts\") pod \"2bd79f24-e81e-49bd-917d-5a6e3281f336\" (UID: \"2bd79f24-e81e-49bd-917d-5a6e3281f336\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.523415 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lq2w\" (UniqueName: \"kubernetes.io/projected/75e19b5f-83d0-4a02-8791-4da2749e0a12-kube-api-access-8lq2w\") pod \"75e19b5f-83d0-4a02-8791-4da2749e0a12\" (UID: \"75e19b5f-83d0-4a02-8791-4da2749e0a12\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.524458 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75e19b5f-83d0-4a02-8791-4da2749e0a12-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "75e19b5f-83d0-4a02-8791-4da2749e0a12" (UID: "75e19b5f-83d0-4a02-8791-4da2749e0a12"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.524484 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnd8c\" (UniqueName: \"kubernetes.io/projected/d94a8c9d-5f73-44fa-8fc7-f725f21a8127-kube-api-access-xnd8c\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.524611 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34db090c-22b5-4462-a0ef-aca71a75a21a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34db090c-22b5-4462-a0ef-aca71a75a21a" (UID: "34db090c-22b5-4462-a0ef-aca71a75a21a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.525132 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bd79f24-e81e-49bd-917d-5a6e3281f336-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2bd79f24-e81e-49bd-917d-5a6e3281f336" (UID: "2bd79f24-e81e-49bd-917d-5a6e3281f336"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.537600 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tn2wq" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.538061 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tn2wq" event={"ID":"f692bc07-eca1-4b0b-b5fc-a9ad3aa34232","Type":"ContainerDied","Data":"c21c9674ef4096b737baf72b4a540f8f36a04aa83bf95816d75f616b50a0a26c"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.538167 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c21c9674ef4096b737baf72b4a540f8f36a04aa83bf95816d75f616b50a0a26c" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.538244 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e19b5f-83d0-4a02-8791-4da2749e0a12-kube-api-access-8lq2w" (OuterVolumeSpecName: "kube-api-access-8lq2w") pod "75e19b5f-83d0-4a02-8791-4da2749e0a12" (UID: "75e19b5f-83d0-4a02-8791-4da2749e0a12"). InnerVolumeSpecName "kube-api-access-8lq2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.540385 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd79f24-e81e-49bd-917d-5a6e3281f336-kube-api-access-wt75j" (OuterVolumeSpecName: "kube-api-access-wt75j") pod "2bd79f24-e81e-49bd-917d-5a6e3281f336" (UID: "2bd79f24-e81e-49bd-917d-5a6e3281f336"). InnerVolumeSpecName "kube-api-access-wt75j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.542208 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34db090c-22b5-4462-a0ef-aca71a75a21a-kube-api-access-thtvp" (OuterVolumeSpecName: "kube-api-access-thtvp") pod "34db090c-22b5-4462-a0ef-aca71a75a21a" (UID: "34db090c-22b5-4462-a0ef-aca71a75a21a"). InnerVolumeSpecName "kube-api-access-thtvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.551252 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pk76z" event={"ID":"fcc6ad68-0f83-4a73-b83a-46eca178604e","Type":"ContainerDied","Data":"107b49ae018da26dbe839a46e23c34ee287b9b6f89fc88957b3d710bd2071b45"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.551291 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="107b49ae018da26dbe839a46e23c34ee287b9b6f89fc88957b3d710bd2071b45" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.551340 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pk76z" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.589162 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5008baa-3c8b-4048-a87e-4e6f3e5d2327","Type":"ContainerStarted","Data":"8fa23b941ff47b4c640c049f22cfc922b69622ffeb59f2719b81800fcc49f39f"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.596964 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c33bbab-5535-4d34-8b87-90c7e1549281","Type":"ContainerStarted","Data":"c59fedf70eefd7ee5cf0a6d98b48689ed03003bd111e68ef5790920a60d8c340"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.607387 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-caff-account-create-update-sgpcw" event={"ID":"2bd79f24-e81e-49bd-917d-5a6e3281f336","Type":"ContainerDied","Data":"1ae3300fa5c2605e28ea6fe281136d4ad37dd1b115ffb6f578de0f55fab4a804"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.607431 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ae3300fa5c2605e28ea6fe281136d4ad37dd1b115ffb6f578de0f55fab4a804" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.607488 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-caff-account-create-update-sgpcw" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.629295 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fcc6ad68-0f83-4a73-b83a-46eca178604e-operator-scripts\") pod \"fcc6ad68-0f83-4a73-b83a-46eca178604e\" (UID: \"fcc6ad68-0f83-4a73-b83a-46eca178604e\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.629487 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlw88\" (UniqueName: \"kubernetes.io/projected/fcc6ad68-0f83-4a73-b83a-46eca178604e-kube-api-access-xlw88\") pod \"fcc6ad68-0f83-4a73-b83a-46eca178604e\" (UID: \"fcc6ad68-0f83-4a73-b83a-46eca178604e\") " Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.629990 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vlctk" event={"ID":"c0f69f89-4463-4f8a-b892-ca81630f4b67","Type":"ContainerDied","Data":"038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.630220 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcc6ad68-0f83-4a73-b83a-46eca178604e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fcc6ad68-0f83-4a73-b83a-46eca178604e" (UID: "fcc6ad68-0f83-4a73-b83a-46eca178604e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.630359 4848 generic.go:334] "Generic (PLEG): container finished" podID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerID="038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785" exitCode=0 Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.634328 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fcc6ad68-0f83-4a73-b83a-46eca178604e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.634484 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75e19b5f-83d0-4a02-8791-4da2749e0a12-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.637897 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34db090c-22b5-4462-a0ef-aca71a75a21a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.637932 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt75j\" (UniqueName: \"kubernetes.io/projected/2bd79f24-e81e-49bd-917d-5a6e3281f336-kube-api-access-wt75j\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.637966 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thtvp\" (UniqueName: \"kubernetes.io/projected/34db090c-22b5-4462-a0ef-aca71a75a21a-kube-api-access-thtvp\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.637980 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bd79f24-e81e-49bd-917d-5a6e3281f336-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.638287 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcc6ad68-0f83-4a73-b83a-46eca178604e-kube-api-access-xlw88" (OuterVolumeSpecName: "kube-api-access-xlw88") pod "fcc6ad68-0f83-4a73-b83a-46eca178604e" (UID: "fcc6ad68-0f83-4a73-b83a-46eca178604e"). InnerVolumeSpecName "kube-api-access-xlw88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.638590 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lq2w\" (UniqueName: \"kubernetes.io/projected/75e19b5f-83d0-4a02-8791-4da2749e0a12-kube-api-access-8lq2w\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.645183 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3873-account-create-update-nc5n8" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.646016 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3873-account-create-update-nc5n8" event={"ID":"34db090c-22b5-4462-a0ef-aca71a75a21a","Type":"ContainerDied","Data":"ea2f66ea4dd001e26b15e3525988cf1b5ea400a47b0835c20da0ebdb97143c72"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.646080 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea2f66ea4dd001e26b15e3525988cf1b5ea400a47b0835c20da0ebdb97143c72" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.664468 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.664450835 podStartE2EDuration="5.664450835s" podCreationTimestamp="2025-12-04 14:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:37.642519599 +0000 UTC m=+1521.585016127" watchObservedRunningTime="2025-12-04 14:13:37.664450835 +0000 UTC m=+1521.606947363" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.669713 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerStarted","Data":"d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.675617 4848 scope.go:117] "RemoveContainer" containerID="c4f0e98194b7e03b78802853c99d8014a587b5764feac053e0b254bdcc6c066a" Dec 04 14:13:37 crc kubenswrapper[4848]: E1204 14:13:37.676669 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-878c6546-2c98w_openstack(a9478f9e-a08b-4d95-b57a-fb8ed97850fb)\"" pod="openstack/heat-api-878c6546-2c98w" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.676742 4848 scope.go:117] "RemoveContainer" containerID="85e1878ee8070715c8ff9e255a06174863bb49f388e2ebceb783c5bef7f70b5b" Dec 04 14:13:37 crc kubenswrapper[4848]: E1204 14:13:37.676922 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58569cf956-f89kq_openstack(28dda188-1082-4b1b-a0dc-42297a0918b4)\"" pod="openstack/heat-cfnapi-58569cf956-f89kq" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.680108 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6fc5j" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.683073 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6fc5j" event={"ID":"75e19b5f-83d0-4a02-8791-4da2749e0a12","Type":"ContainerDied","Data":"3b41e6d4d8cffe93315d37215100bd1db5ff27556ce1f8d5cf1fd5428183b375"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.683127 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b41e6d4d8cffe93315d37215100bd1db5ff27556ce1f8d5cf1fd5428183b375" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.693492 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e824-account-create-update-h466m" event={"ID":"d94a8c9d-5f73-44fa-8fc7-f725f21a8127","Type":"ContainerDied","Data":"93ae6e3ee3c4da480f880422b4ba0ee42dfac72536f71893de57fc053da54901"} Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.693530 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93ae6e3ee3c4da480f880422b4ba0ee42dfac72536f71893de57fc053da54901" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.693580 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e824-account-create-update-h466m" Dec 04 14:13:37 crc kubenswrapper[4848]: I1204 14:13:37.749333 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlw88\" (UniqueName: \"kubernetes.io/projected/fcc6ad68-0f83-4a73-b83a-46eca178604e-kube-api-access-xlw88\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.406558 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b349db-6f99-4351-8a39-7dc1ba31496d" path="/var/lib/kubelet/pods/88b349db-6f99-4351-8a39-7dc1ba31496d/volumes" Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.704662 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerStarted","Data":"484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97"} Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.705119 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.704814 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="proxy-httpd" containerID="cri-o://484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97" gracePeriod=30 Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.704877 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="sg-core" containerID="cri-o://d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee" gracePeriod=30 Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.704852 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="ceilometer-notification-agent" containerID="cri-o://d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c" gracePeriod=30 Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.704757 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="ceilometer-central-agent" containerID="cri-o://564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6" gracePeriod=30 Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.706880 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vlctk" event={"ID":"c0f69f89-4463-4f8a-b892-ca81630f4b67","Type":"ContainerStarted","Data":"be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56"} Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.713454 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c33bbab-5535-4d34-8b87-90c7e1549281","Type":"ContainerStarted","Data":"ec6b38b4f85cddacd3c928e4f7755058c432ed1ec15e5ba63c536c8b9ec6423c"} Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.713490 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0c33bbab-5535-4d34-8b87-90c7e1549281","Type":"ContainerStarted","Data":"0e3a975e195edf3a6a082166f0436585ca2ca3ccb6e7cce9c6714543c3102c03"} Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.714365 4848 scope.go:117] "RemoveContainer" containerID="85e1878ee8070715c8ff9e255a06174863bb49f388e2ebceb783c5bef7f70b5b" Dec 04 14:13:38 crc kubenswrapper[4848]: E1204 14:13:38.714651 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58569cf956-f89kq_openstack(28dda188-1082-4b1b-a0dc-42297a0918b4)\"" pod="openstack/heat-cfnapi-58569cf956-f89kq" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.714711 4848 scope.go:117] "RemoveContainer" containerID="c4f0e98194b7e03b78802853c99d8014a587b5764feac053e0b254bdcc6c066a" Dec 04 14:13:38 crc kubenswrapper[4848]: E1204 14:13:38.715018 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-878c6546-2c98w_openstack(a9478f9e-a08b-4d95-b57a-fb8ed97850fb)\"" pod="openstack/heat-api-878c6546-2c98w" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.743777 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.209351025 podStartE2EDuration="11.743760741s" podCreationTimestamp="2025-12-04 14:13:27 +0000 UTC" firstStartedPulling="2025-12-04 14:13:32.054683536 +0000 UTC m=+1515.997180064" lastFinishedPulling="2025-12-04 14:13:37.589093252 +0000 UTC m=+1521.531589780" observedRunningTime="2025-12-04 14:13:38.738623845 +0000 UTC m=+1522.681120373" watchObservedRunningTime="2025-12-04 14:13:38.743760741 +0000 UTC m=+1522.686257269" Dec 04 14:13:38 crc kubenswrapper[4848]: I1204 14:13:38.792996 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.792957133 podStartE2EDuration="4.792957133s" podCreationTimestamp="2025-12-04 14:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:13:38.778112157 +0000 UTC m=+1522.720608675" watchObservedRunningTime="2025-12-04 14:13:38.792957133 +0000 UTC m=+1522.735453651" Dec 04 14:13:39 crc kubenswrapper[4848]: I1204 14:13:39.729912 4848 generic.go:334] "Generic (PLEG): container finished" podID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerID="be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56" exitCode=0 Dec 04 14:13:39 crc kubenswrapper[4848]: I1204 14:13:39.730079 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vlctk" event={"ID":"c0f69f89-4463-4f8a-b892-ca81630f4b67","Type":"ContainerDied","Data":"be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56"} Dec 04 14:13:39 crc kubenswrapper[4848]: I1204 14:13:39.740931 4848 generic.go:334] "Generic (PLEG): container finished" podID="47dc56bf-c795-49b1-ad78-c8407438f671" containerID="484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97" exitCode=0 Dec 04 14:13:39 crc kubenswrapper[4848]: I1204 14:13:39.741196 4848 generic.go:334] "Generic (PLEG): container finished" podID="47dc56bf-c795-49b1-ad78-c8407438f671" containerID="d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee" exitCode=2 Dec 04 14:13:39 crc kubenswrapper[4848]: I1204 14:13:39.741210 4848 generic.go:334] "Generic (PLEG): container finished" podID="47dc56bf-c795-49b1-ad78-c8407438f671" containerID="d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c" exitCode=0 Dec 04 14:13:39 crc kubenswrapper[4848]: I1204 14:13:39.742797 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerDied","Data":"484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97"} Dec 04 14:13:39 crc kubenswrapper[4848]: I1204 14:13:39.742840 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerDied","Data":"d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee"} Dec 04 14:13:39 crc kubenswrapper[4848]: I1204 14:13:39.742853 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerDied","Data":"d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c"} Dec 04 14:13:40 crc kubenswrapper[4848]: I1204 14:13:40.044570 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406012 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4czxk"] Dec 04 14:13:41 crc kubenswrapper[4848]: E1204 14:13:41.406601 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b349db-6f99-4351-8a39-7dc1ba31496d" containerName="dnsmasq-dns" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406617 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b349db-6f99-4351-8a39-7dc1ba31496d" containerName="dnsmasq-dns" Dec 04 14:13:41 crc kubenswrapper[4848]: E1204 14:13:41.406633 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e19b5f-83d0-4a02-8791-4da2749e0a12" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406641 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e19b5f-83d0-4a02-8791-4da2749e0a12" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: E1204 14:13:41.406661 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f692bc07-eca1-4b0b-b5fc-a9ad3aa34232" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406669 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f692bc07-eca1-4b0b-b5fc-a9ad3aa34232" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: E1204 14:13:41.406688 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34db090c-22b5-4462-a0ef-aca71a75a21a" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406695 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="34db090c-22b5-4462-a0ef-aca71a75a21a" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: E1204 14:13:41.406726 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc6ad68-0f83-4a73-b83a-46eca178604e" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406733 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc6ad68-0f83-4a73-b83a-46eca178604e" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: E1204 14:13:41.406748 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d94a8c9d-5f73-44fa-8fc7-f725f21a8127" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406755 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d94a8c9d-5f73-44fa-8fc7-f725f21a8127" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: E1204 14:13:41.406774 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b349db-6f99-4351-8a39-7dc1ba31496d" containerName="init" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406780 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b349db-6f99-4351-8a39-7dc1ba31496d" containerName="init" Dec 04 14:13:41 crc kubenswrapper[4848]: E1204 14:13:41.406794 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd79f24-e81e-49bd-917d-5a6e3281f336" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.406801 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd79f24-e81e-49bd-917d-5a6e3281f336" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.407055 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd79f24-e81e-49bd-917d-5a6e3281f336" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.407073 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d94a8c9d-5f73-44fa-8fc7-f725f21a8127" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.407086 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e19b5f-83d0-4a02-8791-4da2749e0a12" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.407099 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f692bc07-eca1-4b0b-b5fc-a9ad3aa34232" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.407110 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b349db-6f99-4351-8a39-7dc1ba31496d" containerName="dnsmasq-dns" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.407121 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="34db090c-22b5-4462-a0ef-aca71a75a21a" containerName="mariadb-account-create-update" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.407135 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc6ad68-0f83-4a73-b83a-46eca178604e" containerName="mariadb-database-create" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.408265 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.415284 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-48wpm" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.415305 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.415501 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.416419 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4czxk"] Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.469707 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-config-data\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.469802 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.469998 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-scripts\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.470115 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk5bt\" (UniqueName: \"kubernetes.io/projected/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-kube-api-access-sk5bt\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.572138 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk5bt\" (UniqueName: \"kubernetes.io/projected/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-kube-api-access-sk5bt\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.572188 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-config-data\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.572240 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.572378 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-scripts\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.581449 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-scripts\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.583546 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-config-data\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.590787 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.591906 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk5bt\" (UniqueName: \"kubernetes.io/projected/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-kube-api-access-sk5bt\") pod \"nova-cell0-conductor-db-sync-4czxk\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:41 crc kubenswrapper[4848]: I1204 14:13:41.735072 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.018403 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vlctk" event={"ID":"c0f69f89-4463-4f8a-b892-ca81630f4b67","Type":"ContainerStarted","Data":"a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8"} Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.045236 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vlctk" podStartSLOduration=4.68223991 podStartE2EDuration="8.045215924s" podCreationTimestamp="2025-12-04 14:13:34 +0000 UTC" firstStartedPulling="2025-12-04 14:13:37.631419868 +0000 UTC m=+1521.573916406" lastFinishedPulling="2025-12-04 14:13:40.994395892 +0000 UTC m=+1524.936892420" observedRunningTime="2025-12-04 14:13:42.037653013 +0000 UTC m=+1525.980149541" watchObservedRunningTime="2025-12-04 14:13:42.045215924 +0000 UTC m=+1525.987712452" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.095402 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.157081 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-878c6546-2c98w"] Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.380319 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4czxk"] Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.389361 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:13:42 crc kubenswrapper[4848]: W1204 14:13:42.422891 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod013c9f72_2dcd_4e71_a4ab_aec59a8131aa.slice/crio-6327a5afb22da183a566e78ab5cd15137d2fc7609e40e888f97b6dc2e0845cab WatchSource:0}: Error finding container 6327a5afb22da183a566e78ab5cd15137d2fc7609e40e888f97b6dc2e0845cab: Status 404 returned error can't find the container with id 6327a5afb22da183a566e78ab5cd15137d2fc7609e40e888f97b6dc2e0845cab Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.483092 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-58569cf956-f89kq"] Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.595458 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.752543 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.812914 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data-custom\") pod \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.813127 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-combined-ca-bundle\") pod \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.813192 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data\") pod \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.813275 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4tj9\" (UniqueName: \"kubernetes.io/projected/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-kube-api-access-g4tj9\") pod \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\" (UID: \"a9478f9e-a08b-4d95-b57a-fb8ed97850fb\") " Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.822247 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-kube-api-access-g4tj9" (OuterVolumeSpecName: "kube-api-access-g4tj9") pod "a9478f9e-a08b-4d95-b57a-fb8ed97850fb" (UID: "a9478f9e-a08b-4d95-b57a-fb8ed97850fb"). InnerVolumeSpecName "kube-api-access-g4tj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.851137 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a9478f9e-a08b-4d95-b57a-fb8ed97850fb" (UID: "a9478f9e-a08b-4d95-b57a-fb8ed97850fb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.878039 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9478f9e-a08b-4d95-b57a-fb8ed97850fb" (UID: "a9478f9e-a08b-4d95-b57a-fb8ed97850fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.915762 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4tj9\" (UniqueName: \"kubernetes.io/projected/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-kube-api-access-g4tj9\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.915795 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.915803 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.957937 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:42 crc kubenswrapper[4848]: I1204 14:13:42.960126 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data" (OuterVolumeSpecName: "config-data") pod "a9478f9e-a08b-4d95-b57a-fb8ed97850fb" (UID: "a9478f9e-a08b-4d95-b57a-fb8ed97850fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.017013 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data-custom\") pod \"28dda188-1082-4b1b-a0dc-42297a0918b4\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.017925 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/28dda188-1082-4b1b-a0dc-42297a0918b4-kube-api-access-cdcsz\") pod \"28dda188-1082-4b1b-a0dc-42297a0918b4\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.018458 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data\") pod \"28dda188-1082-4b1b-a0dc-42297a0918b4\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.018710 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-combined-ca-bundle\") pod \"28dda188-1082-4b1b-a0dc-42297a0918b4\" (UID: \"28dda188-1082-4b1b-a0dc-42297a0918b4\") " Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.019546 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9478f9e-a08b-4d95-b57a-fb8ed97850fb-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.022729 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28dda188-1082-4b1b-a0dc-42297a0918b4-kube-api-access-cdcsz" (OuterVolumeSpecName: "kube-api-access-cdcsz") pod "28dda188-1082-4b1b-a0dc-42297a0918b4" (UID: "28dda188-1082-4b1b-a0dc-42297a0918b4"). InnerVolumeSpecName "kube-api-access-cdcsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.022797 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "28dda188-1082-4b1b-a0dc-42297a0918b4" (UID: "28dda188-1082-4b1b-a0dc-42297a0918b4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.036690 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58569cf956-f89kq" event={"ID":"28dda188-1082-4b1b-a0dc-42297a0918b4","Type":"ContainerDied","Data":"cb3e395fc1e3f68a5acfaa67d87dfabe72743efc0e5c614e2400a7e38283fa35"} Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.036744 4848 scope.go:117] "RemoveContainer" containerID="85e1878ee8070715c8ff9e255a06174863bb49f388e2ebceb783c5bef7f70b5b" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.036825 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58569cf956-f89kq" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.046071 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-878c6546-2c98w" event={"ID":"a9478f9e-a08b-4d95-b57a-fb8ed97850fb","Type":"ContainerDied","Data":"d7e28a1e8e7ee8c99e9d551eeaf2e177f76306841e0b48c0be27bd0a8a8cb657"} Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.046130 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-878c6546-2c98w" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.051207 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4czxk" event={"ID":"013c9f72-2dcd-4e71-a4ab-aec59a8131aa","Type":"ContainerStarted","Data":"6327a5afb22da183a566e78ab5cd15137d2fc7609e40e888f97b6dc2e0845cab"} Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.073211 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28dda188-1082-4b1b-a0dc-42297a0918b4" (UID: "28dda188-1082-4b1b-a0dc-42297a0918b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.087507 4848 scope.go:117] "RemoveContainer" containerID="c4f0e98194b7e03b78802853c99d8014a587b5764feac053e0b254bdcc6c066a" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.104579 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-878c6546-2c98w"] Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.120145 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-878c6546-2c98w"] Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.120862 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data" (OuterVolumeSpecName: "config-data") pod "28dda188-1082-4b1b-a0dc-42297a0918b4" (UID: "28dda188-1082-4b1b-a0dc-42297a0918b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.121469 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.121499 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.121510 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28dda188-1082-4b1b-a0dc-42297a0918b4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.121520 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/28dda188-1082-4b1b-a0dc-42297a0918b4-kube-api-access-cdcsz\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.210059 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.377884 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-58569cf956-f89kq"] Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.389065 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-58569cf956-f89kq"] Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.433662 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.434226 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.489132 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 14:13:43 crc kubenswrapper[4848]: I1204 14:13:43.491175 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 14:13:44 crc kubenswrapper[4848]: I1204 14:13:44.065746 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:13:44 crc kubenswrapper[4848]: I1204 14:13:44.066170 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 14:13:44 crc kubenswrapper[4848]: I1204 14:13:44.406648 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" path="/var/lib/kubelet/pods/28dda188-1082-4b1b-a0dc-42297a0918b4/volumes" Dec 04 14:13:44 crc kubenswrapper[4848]: I1204 14:13:44.407668 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" path="/var/lib/kubelet/pods/a9478f9e-a08b-4d95-b57a-fb8ed97850fb/volumes" Dec 04 14:13:44 crc kubenswrapper[4848]: I1204 14:13:44.798813 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:44 crc kubenswrapper[4848]: I1204 14:13:44.799112 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:44 crc kubenswrapper[4848]: I1204 14:13:44.886370 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:45 crc kubenswrapper[4848]: I1204 14:13:45.618124 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:45 crc kubenswrapper[4848]: I1204 14:13:45.618370 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:45 crc kubenswrapper[4848]: I1204 14:13:45.658168 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:45 crc kubenswrapper[4848]: I1204 14:13:45.684065 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.099883 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.118372 4848 generic.go:334] "Generic (PLEG): container finished" podID="47dc56bf-c795-49b1-ad78-c8407438f671" containerID="564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6" exitCode=0 Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.120070 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerDied","Data":"564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6"} Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.120144 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47dc56bf-c795-49b1-ad78-c8407438f671","Type":"ContainerDied","Data":"bd2741cbb4367c200745e3e5a145842265cf81aa8b51d8c6bef6cb392c5e6873"} Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.120173 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.120195 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.120214 4848 scope.go:117] "RemoveContainer" containerID="484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.120255 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.175517 4848 scope.go:117] "RemoveContainer" containerID="d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.244236 4848 scope.go:117] "RemoveContainer" containerID="d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.291857 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb248\" (UniqueName: \"kubernetes.io/projected/47dc56bf-c795-49b1-ad78-c8407438f671-kube-api-access-sb248\") pod \"47dc56bf-c795-49b1-ad78-c8407438f671\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.291924 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-combined-ca-bundle\") pod \"47dc56bf-c795-49b1-ad78-c8407438f671\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.292046 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-run-httpd\") pod \"47dc56bf-c795-49b1-ad78-c8407438f671\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.292156 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-log-httpd\") pod \"47dc56bf-c795-49b1-ad78-c8407438f671\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.292193 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-sg-core-conf-yaml\") pod \"47dc56bf-c795-49b1-ad78-c8407438f671\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.292275 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-scripts\") pod \"47dc56bf-c795-49b1-ad78-c8407438f671\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.292310 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-config-data\") pod \"47dc56bf-c795-49b1-ad78-c8407438f671\" (UID: \"47dc56bf-c795-49b1-ad78-c8407438f671\") " Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.293102 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "47dc56bf-c795-49b1-ad78-c8407438f671" (UID: "47dc56bf-c795-49b1-ad78-c8407438f671"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.293280 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "47dc56bf-c795-49b1-ad78-c8407438f671" (UID: "47dc56bf-c795-49b1-ad78-c8407438f671"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.296662 4848 scope.go:117] "RemoveContainer" containerID="564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.302299 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47dc56bf-c795-49b1-ad78-c8407438f671-kube-api-access-sb248" (OuterVolumeSpecName: "kube-api-access-sb248") pod "47dc56bf-c795-49b1-ad78-c8407438f671" (UID: "47dc56bf-c795-49b1-ad78-c8407438f671"). InnerVolumeSpecName "kube-api-access-sb248". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.318795 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-scripts" (OuterVolumeSpecName: "scripts") pod "47dc56bf-c795-49b1-ad78-c8407438f671" (UID: "47dc56bf-c795-49b1-ad78-c8407438f671"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.383199 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "47dc56bf-c795-49b1-ad78-c8407438f671" (UID: "47dc56bf-c795-49b1-ad78-c8407438f671"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.395292 4848 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.395326 4848 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47dc56bf-c795-49b1-ad78-c8407438f671-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.395339 4848 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.395351 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.395364 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb248\" (UniqueName: \"kubernetes.io/projected/47dc56bf-c795-49b1-ad78-c8407438f671-kube-api-access-sb248\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.463696 4848 scope.go:117] "RemoveContainer" containerID="484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.465093 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97\": container with ID starting with 484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97 not found: ID does not exist" containerID="484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.465160 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97"} err="failed to get container status \"484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97\": rpc error: code = NotFound desc = could not find container \"484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97\": container with ID starting with 484d2ef8b2ebda11cd36858c56a6e170d181147683d5120997ac64863a78fb97 not found: ID does not exist" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.465193 4848 scope.go:117] "RemoveContainer" containerID="d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.474587 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee\": container with ID starting with d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee not found: ID does not exist" containerID="d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.475164 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee"} err="failed to get container status \"d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee\": rpc error: code = NotFound desc = could not find container \"d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee\": container with ID starting with d3d9895211b00a3a3539c5ae1ab9005727bf6eed6dba8d96422fe886cfb113ee not found: ID does not exist" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.475273 4848 scope.go:117] "RemoveContainer" containerID="d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.487390 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c\": container with ID starting with d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c not found: ID does not exist" containerID="d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.487448 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c"} err="failed to get container status \"d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c\": rpc error: code = NotFound desc = could not find container \"d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c\": container with ID starting with d132e04308e24a79ff602f6daaeb758d130c21f6554b1ae12c1a437900ea3f9c not found: ID does not exist" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.487481 4848 scope.go:117] "RemoveContainer" containerID="564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.488942 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6\": container with ID starting with 564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6 not found: ID does not exist" containerID="564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.489023 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6"} err="failed to get container status \"564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6\": rpc error: code = NotFound desc = could not find container \"564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6\": container with ID starting with 564771b5b92436cd99be4e9caefedeb565efa76492ab606d18e1ac727051b6c6 not found: ID does not exist" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.549151 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47dc56bf-c795-49b1-ad78-c8407438f671" (UID: "47dc56bf-c795-49b1-ad78-c8407438f671"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.587410 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-config-data" (OuterVolumeSpecName: "config-data") pod "47dc56bf-c795-49b1-ad78-c8407438f671" (UID: "47dc56bf-c795-49b1-ad78-c8407438f671"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.603447 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.603495 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47dc56bf-c795-49b1-ad78-c8407438f671-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.614262 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.727387 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7467697848-tlr74"] Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.727831 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-7467697848-tlr74" podUID="ca5459c4-87fe-414e-9e6e-843d0e6329a9" containerName="heat-engine" containerID="cri-o://7558f74fc22966b329e7619df76d7df2a8dcb0d46671e858ea57020159bf533b" gracePeriod=60 Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.783006 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.808470 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.835351 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.835878 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="proxy-httpd" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.835897 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="proxy-httpd" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.835916 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="ceilometer-notification-agent" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.835923 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="ceilometer-notification-agent" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.835941 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" containerName="heat-api" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.835961 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" containerName="heat-api" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.835989 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" containerName="heat-cfnapi" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.835995 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" containerName="heat-cfnapi" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.836014 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="ceilometer-central-agent" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836020 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="ceilometer-central-agent" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.836029 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" containerName="heat-cfnapi" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836036 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" containerName="heat-cfnapi" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.836049 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="sg-core" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836054 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="sg-core" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836272 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" containerName="heat-cfnapi" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836285 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="sg-core" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836296 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" containerName="heat-api" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836306 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" containerName="heat-api" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836316 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="ceilometer-notification-agent" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836326 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="proxy-httpd" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836340 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" containerName="ceilometer-central-agent" Dec 04 14:13:46 crc kubenswrapper[4848]: E1204 14:13:46.836533 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" containerName="heat-api" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836541 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9478f9e-a08b-4d95-b57a-fb8ed97850fb" containerName="heat-api" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.836773 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="28dda188-1082-4b1b-a0dc-42297a0918b4" containerName="heat-cfnapi" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.839084 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.847185 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.847613 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.858758 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.919357 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-log-httpd\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.919492 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-config-data\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.919523 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq4ch\" (UniqueName: \"kubernetes.io/projected/686edfbb-99ba-4e58-a789-ee673f2f4809-kube-api-access-tq4ch\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.919562 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.919579 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-run-httpd\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.919595 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:46 crc kubenswrapper[4848]: I1204 14:13:46.919687 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-scripts\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.021354 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq4ch\" (UniqueName: \"kubernetes.io/projected/686edfbb-99ba-4e58-a789-ee673f2f4809-kube-api-access-tq4ch\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.021422 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.021445 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-run-httpd\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.021461 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.021546 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-scripts\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.021613 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-log-httpd\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.021708 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-config-data\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.022400 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-log-httpd\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.022637 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-run-httpd\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.026085 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-scripts\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.026284 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.029700 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-config-data\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.033333 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.068993 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq4ch\" (UniqueName: \"kubernetes.io/projected/686edfbb-99ba-4e58-a789-ee673f2f4809-kube-api-access-tq4ch\") pod \"ceilometer-0\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " pod="openstack/ceilometer-0" Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.120269 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:47 crc kubenswrapper[4848]: I1204 14:13:47.121154 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:13:48 crc kubenswrapper[4848]: I1204 14:13:48.415906 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47dc56bf-c795-49b1-ad78-c8407438f671" path="/var/lib/kubelet/pods/47dc56bf-c795-49b1-ad78-c8407438f671/volumes" Dec 04 14:13:48 crc kubenswrapper[4848]: I1204 14:13:48.708814 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 14:13:48 crc kubenswrapper[4848]: I1204 14:13:48.709022 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 14:13:48 crc kubenswrapper[4848]: I1204 14:13:48.720356 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 14:13:48 crc kubenswrapper[4848]: I1204 14:13:48.731435 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:48 crc kubenswrapper[4848]: I1204 14:13:48.731564 4848 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 14:13:48 crc kubenswrapper[4848]: I1204 14:13:48.734628 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 14:13:49 crc kubenswrapper[4848]: E1204 14:13:49.977092 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7558f74fc22966b329e7619df76d7df2a8dcb0d46671e858ea57020159bf533b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:13:49 crc kubenswrapper[4848]: E1204 14:13:49.979442 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7558f74fc22966b329e7619df76d7df2a8dcb0d46671e858ea57020159bf533b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:13:49 crc kubenswrapper[4848]: E1204 14:13:49.981045 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7558f74fc22966b329e7619df76d7df2a8dcb0d46671e858ea57020159bf533b" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:13:49 crc kubenswrapper[4848]: E1204 14:13:49.981106 4848 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7467697848-tlr74" podUID="ca5459c4-87fe-414e-9e6e-843d0e6329a9" containerName="heat-engine" Dec 04 14:13:54 crc kubenswrapper[4848]: I1204 14:13:54.874227 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:54 crc kubenswrapper[4848]: I1204 14:13:54.947564 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vlctk"] Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.274149 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vlctk" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerName="registry-server" containerID="cri-o://a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8" gracePeriod=2 Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.274314 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4czxk" event={"ID":"013c9f72-2dcd-4e71-a4ab-aec59a8131aa","Type":"ContainerStarted","Data":"24bd7e7b81eb628112fa037f245a03d4b2f307ab5c11fb2a7de3b4013b9afa74"} Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.294793 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-4czxk" podStartSLOduration=1.83520212 podStartE2EDuration="14.294774098s" podCreationTimestamp="2025-12-04 14:13:41 +0000 UTC" firstStartedPulling="2025-12-04 14:13:42.428528618 +0000 UTC m=+1526.371025146" lastFinishedPulling="2025-12-04 14:13:54.888100596 +0000 UTC m=+1538.830597124" observedRunningTime="2025-12-04 14:13:55.292804734 +0000 UTC m=+1539.235301262" watchObservedRunningTime="2025-12-04 14:13:55.294774098 +0000 UTC m=+1539.237270626" Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.483557 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:13:55 crc kubenswrapper[4848]: W1204 14:13:55.483639 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod686edfbb_99ba_4e58_a789_ee673f2f4809.slice/crio-ffa91062ce86e11b50d5f95087bff6406ad36bc0923c9522663c49043f168ff7 WatchSource:0}: Error finding container ffa91062ce86e11b50d5f95087bff6406ad36bc0923c9522663c49043f168ff7: Status 404 returned error can't find the container with id ffa91062ce86e11b50d5f95087bff6406ad36bc0923c9522663c49043f168ff7 Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.832542 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.871087 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-catalog-content\") pod \"c0f69f89-4463-4f8a-b892-ca81630f4b67\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.875237 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-utilities\") pod \"c0f69f89-4463-4f8a-b892-ca81630f4b67\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.875303 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkbxw\" (UniqueName: \"kubernetes.io/projected/c0f69f89-4463-4f8a-b892-ca81630f4b67-kube-api-access-kkbxw\") pod \"c0f69f89-4463-4f8a-b892-ca81630f4b67\" (UID: \"c0f69f89-4463-4f8a-b892-ca81630f4b67\") " Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.875799 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-utilities" (OuterVolumeSpecName: "utilities") pod "c0f69f89-4463-4f8a-b892-ca81630f4b67" (UID: "c0f69f89-4463-4f8a-b892-ca81630f4b67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.876434 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.881099 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0f69f89-4463-4f8a-b892-ca81630f4b67-kube-api-access-kkbxw" (OuterVolumeSpecName: "kube-api-access-kkbxw") pod "c0f69f89-4463-4f8a-b892-ca81630f4b67" (UID: "c0f69f89-4463-4f8a-b892-ca81630f4b67"). InnerVolumeSpecName "kube-api-access-kkbxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.951927 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0f69f89-4463-4f8a-b892-ca81630f4b67" (UID: "c0f69f89-4463-4f8a-b892-ca81630f4b67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.978145 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0f69f89-4463-4f8a-b892-ca81630f4b67-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:55 crc kubenswrapper[4848]: I1204 14:13:55.978176 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkbxw\" (UniqueName: \"kubernetes.io/projected/c0f69f89-4463-4f8a-b892-ca81630f4b67-kube-api-access-kkbxw\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.312271 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerStarted","Data":"4df367335a2f8aa56f458612112434f462be68af1da23da029c11041c9d1ce39"} Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.312874 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerStarted","Data":"ffa91062ce86e11b50d5f95087bff6406ad36bc0923c9522663c49043f168ff7"} Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.326412 4848 generic.go:334] "Generic (PLEG): container finished" podID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerID="a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8" exitCode=0 Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.326647 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vlctk" event={"ID":"c0f69f89-4463-4f8a-b892-ca81630f4b67","Type":"ContainerDied","Data":"a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8"} Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.326733 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vlctk" event={"ID":"c0f69f89-4463-4f8a-b892-ca81630f4b67","Type":"ContainerDied","Data":"c30c9472e3d5249c219f4d2aebf369e794d2725804a0d1d95e4bcea870923fd0"} Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.326796 4848 scope.go:117] "RemoveContainer" containerID="a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.326980 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vlctk" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.343234 4848 generic.go:334] "Generic (PLEG): container finished" podID="ca5459c4-87fe-414e-9e6e-843d0e6329a9" containerID="7558f74fc22966b329e7619df76d7df2a8dcb0d46671e858ea57020159bf533b" exitCode=0 Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.343539 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7467697848-tlr74" event={"ID":"ca5459c4-87fe-414e-9e6e-843d0e6329a9","Type":"ContainerDied","Data":"7558f74fc22966b329e7619df76d7df2a8dcb0d46671e858ea57020159bf533b"} Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.371261 4848 scope.go:117] "RemoveContainer" containerID="be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.418791 4848 scope.go:117] "RemoveContainer" containerID="038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.441119 4848 scope.go:117] "RemoveContainer" containerID="a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8" Dec 04 14:13:56 crc kubenswrapper[4848]: E1204 14:13:56.441522 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8\": container with ID starting with a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8 not found: ID does not exist" containerID="a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.441553 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8"} err="failed to get container status \"a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8\": rpc error: code = NotFound desc = could not find container \"a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8\": container with ID starting with a7edd6d1f0f4c7781d25b38d4d7ccf4bc0fe5e2997e25a31e8806f2cf909bdd8 not found: ID does not exist" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.441572 4848 scope.go:117] "RemoveContainer" containerID="be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56" Dec 04 14:13:56 crc kubenswrapper[4848]: E1204 14:13:56.441936 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56\": container with ID starting with be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56 not found: ID does not exist" containerID="be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.441977 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56"} err="failed to get container status \"be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56\": rpc error: code = NotFound desc = could not find container \"be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56\": container with ID starting with be648555c4921530a02a905da934b92dd629151bdc0c81e74848856782b97c56 not found: ID does not exist" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.441990 4848 scope.go:117] "RemoveContainer" containerID="038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785" Dec 04 14:13:56 crc kubenswrapper[4848]: E1204 14:13:56.442729 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785\": container with ID starting with 038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785 not found: ID does not exist" containerID="038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.442754 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785"} err="failed to get container status \"038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785\": rpc error: code = NotFound desc = could not find container \"038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785\": container with ID starting with 038fe7d88dbb9ec6940c4eb023a0d54baba79dffb425b902a0d7103385153785 not found: ID does not exist" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.451240 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vlctk"] Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.451311 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vlctk"] Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.470888 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.604505 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-combined-ca-bundle\") pod \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.604742 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2srvl\" (UniqueName: \"kubernetes.io/projected/ca5459c4-87fe-414e-9e6e-843d0e6329a9-kube-api-access-2srvl\") pod \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.604972 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data\") pod \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.605023 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data-custom\") pod \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\" (UID: \"ca5459c4-87fe-414e-9e6e-843d0e6329a9\") " Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.612815 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca5459c4-87fe-414e-9e6e-843d0e6329a9-kube-api-access-2srvl" (OuterVolumeSpecName: "kube-api-access-2srvl") pod "ca5459c4-87fe-414e-9e6e-843d0e6329a9" (UID: "ca5459c4-87fe-414e-9e6e-843d0e6329a9"). InnerVolumeSpecName "kube-api-access-2srvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.613300 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ca5459c4-87fe-414e-9e6e-843d0e6329a9" (UID: "ca5459c4-87fe-414e-9e6e-843d0e6329a9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.643071 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca5459c4-87fe-414e-9e6e-843d0e6329a9" (UID: "ca5459c4-87fe-414e-9e6e-843d0e6329a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.677407 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data" (OuterVolumeSpecName: "config-data") pod "ca5459c4-87fe-414e-9e6e-843d0e6329a9" (UID: "ca5459c4-87fe-414e-9e6e-843d0e6329a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.707465 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2srvl\" (UniqueName: \"kubernetes.io/projected/ca5459c4-87fe-414e-9e6e-843d0e6329a9-kube-api-access-2srvl\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.707500 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.707509 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:56 crc kubenswrapper[4848]: I1204 14:13:56.707518 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5459c4-87fe-414e-9e6e-843d0e6329a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:57 crc kubenswrapper[4848]: I1204 14:13:57.363297 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerStarted","Data":"f3b4c16fc8876f0840e0e112fa1b0976398995341de26c5f47e141d66e2b620a"} Dec 04 14:13:57 crc kubenswrapper[4848]: I1204 14:13:57.366973 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7467697848-tlr74" event={"ID":"ca5459c4-87fe-414e-9e6e-843d0e6329a9","Type":"ContainerDied","Data":"62fa182304a6d6657dc4a2093ffb6178169940a3a8347e0cf0428624401fc5ea"} Dec 04 14:13:57 crc kubenswrapper[4848]: I1204 14:13:57.367032 4848 scope.go:117] "RemoveContainer" containerID="7558f74fc22966b329e7619df76d7df2a8dcb0d46671e858ea57020159bf533b" Dec 04 14:13:57 crc kubenswrapper[4848]: I1204 14:13:57.367475 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7467697848-tlr74" Dec 04 14:13:57 crc kubenswrapper[4848]: I1204 14:13:57.416110 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7467697848-tlr74"] Dec 04 14:13:57 crc kubenswrapper[4848]: I1204 14:13:57.434694 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-7467697848-tlr74"] Dec 04 14:13:58 crc kubenswrapper[4848]: I1204 14:13:58.382901 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerStarted","Data":"c5998e9cb9c10d8f31bb89d7415262142ac8d10b1b605c8b30129759efb9cedb"} Dec 04 14:13:58 crc kubenswrapper[4848]: I1204 14:13:58.405816 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" path="/var/lib/kubelet/pods/c0f69f89-4463-4f8a-b892-ca81630f4b67/volumes" Dec 04 14:13:58 crc kubenswrapper[4848]: I1204 14:13:58.406781 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca5459c4-87fe-414e-9e6e-843d0e6329a9" path="/var/lib/kubelet/pods/ca5459c4-87fe-414e-9e6e-843d0e6329a9/volumes" Dec 04 14:14:00 crc kubenswrapper[4848]: I1204 14:14:00.420136 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerStarted","Data":"f343cf2906706acd68e0505f577d5274cf8b0fd571cb0aa8908dc95474b2921a"} Dec 04 14:14:00 crc kubenswrapper[4848]: I1204 14:14:00.421175 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:14:00 crc kubenswrapper[4848]: I1204 14:14:00.420912 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="proxy-httpd" containerID="cri-o://f343cf2906706acd68e0505f577d5274cf8b0fd571cb0aa8908dc95474b2921a" gracePeriod=30 Dec 04 14:14:00 crc kubenswrapper[4848]: I1204 14:14:00.420308 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="ceilometer-central-agent" containerID="cri-o://4df367335a2f8aa56f458612112434f462be68af1da23da029c11041c9d1ce39" gracePeriod=30 Dec 04 14:14:00 crc kubenswrapper[4848]: I1204 14:14:00.420935 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="sg-core" containerID="cri-o://c5998e9cb9c10d8f31bb89d7415262142ac8d10b1b605c8b30129759efb9cedb" gracePeriod=30 Dec 04 14:14:00 crc kubenswrapper[4848]: I1204 14:14:00.420971 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="ceilometer-notification-agent" containerID="cri-o://f3b4c16fc8876f0840e0e112fa1b0976398995341de26c5f47e141d66e2b620a" gracePeriod=30 Dec 04 14:14:00 crc kubenswrapper[4848]: I1204 14:14:00.463007 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.93401358 podStartE2EDuration="14.462981267s" podCreationTimestamp="2025-12-04 14:13:46 +0000 UTC" firstStartedPulling="2025-12-04 14:13:55.506125105 +0000 UTC m=+1539.448621623" lastFinishedPulling="2025-12-04 14:13:59.035092782 +0000 UTC m=+1542.977589310" observedRunningTime="2025-12-04 14:14:00.447205149 +0000 UTC m=+1544.389701677" watchObservedRunningTime="2025-12-04 14:14:00.462981267 +0000 UTC m=+1544.405477795" Dec 04 14:14:01 crc kubenswrapper[4848]: I1204 14:14:01.435289 4848 generic.go:334] "Generic (PLEG): container finished" podID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerID="f343cf2906706acd68e0505f577d5274cf8b0fd571cb0aa8908dc95474b2921a" exitCode=0 Dec 04 14:14:01 crc kubenswrapper[4848]: I1204 14:14:01.435579 4848 generic.go:334] "Generic (PLEG): container finished" podID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerID="c5998e9cb9c10d8f31bb89d7415262142ac8d10b1b605c8b30129759efb9cedb" exitCode=2 Dec 04 14:14:01 crc kubenswrapper[4848]: I1204 14:14:01.435587 4848 generic.go:334] "Generic (PLEG): container finished" podID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerID="f3b4c16fc8876f0840e0e112fa1b0976398995341de26c5f47e141d66e2b620a" exitCode=0 Dec 04 14:14:01 crc kubenswrapper[4848]: I1204 14:14:01.435356 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerDied","Data":"f343cf2906706acd68e0505f577d5274cf8b0fd571cb0aa8908dc95474b2921a"} Dec 04 14:14:01 crc kubenswrapper[4848]: I1204 14:14:01.435617 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerDied","Data":"c5998e9cb9c10d8f31bb89d7415262142ac8d10b1b605c8b30129759efb9cedb"} Dec 04 14:14:01 crc kubenswrapper[4848]: I1204 14:14:01.435627 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerDied","Data":"f3b4c16fc8876f0840e0e112fa1b0976398995341de26c5f47e141d66e2b620a"} Dec 04 14:14:10 crc kubenswrapper[4848]: I1204 14:14:10.537518 4848 generic.go:334] "Generic (PLEG): container finished" podID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerID="4df367335a2f8aa56f458612112434f462be68af1da23da029c11041c9d1ce39" exitCode=0 Dec 04 14:14:10 crc kubenswrapper[4848]: I1204 14:14:10.537666 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerDied","Data":"4df367335a2f8aa56f458612112434f462be68af1da23da029c11041c9d1ce39"} Dec 04 14:14:10 crc kubenswrapper[4848]: I1204 14:14:10.545286 4848 generic.go:334] "Generic (PLEG): container finished" podID="013c9f72-2dcd-4e71-a4ab-aec59a8131aa" containerID="24bd7e7b81eb628112fa037f245a03d4b2f307ab5c11fb2a7de3b4013b9afa74" exitCode=0 Dec 04 14:14:10 crc kubenswrapper[4848]: I1204 14:14:10.545329 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4czxk" event={"ID":"013c9f72-2dcd-4e71-a4ab-aec59a8131aa","Type":"ContainerDied","Data":"24bd7e7b81eb628112fa037f245a03d4b2f307ab5c11fb2a7de3b4013b9afa74"} Dec 04 14:14:10 crc kubenswrapper[4848]: I1204 14:14:10.953436 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.097940 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-combined-ca-bundle\") pod \"686edfbb-99ba-4e58-a789-ee673f2f4809\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.098111 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq4ch\" (UniqueName: \"kubernetes.io/projected/686edfbb-99ba-4e58-a789-ee673f2f4809-kube-api-access-tq4ch\") pod \"686edfbb-99ba-4e58-a789-ee673f2f4809\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.098167 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-config-data\") pod \"686edfbb-99ba-4e58-a789-ee673f2f4809\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.098229 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-run-httpd\") pod \"686edfbb-99ba-4e58-a789-ee673f2f4809\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.098311 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-sg-core-conf-yaml\") pod \"686edfbb-99ba-4e58-a789-ee673f2f4809\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.098413 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-scripts\") pod \"686edfbb-99ba-4e58-a789-ee673f2f4809\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.098649 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "686edfbb-99ba-4e58-a789-ee673f2f4809" (UID: "686edfbb-99ba-4e58-a789-ee673f2f4809"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.098822 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-log-httpd\") pod \"686edfbb-99ba-4e58-a789-ee673f2f4809\" (UID: \"686edfbb-99ba-4e58-a789-ee673f2f4809\") " Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.099296 4848 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.099303 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "686edfbb-99ba-4e58-a789-ee673f2f4809" (UID: "686edfbb-99ba-4e58-a789-ee673f2f4809"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.105284 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/686edfbb-99ba-4e58-a789-ee673f2f4809-kube-api-access-tq4ch" (OuterVolumeSpecName: "kube-api-access-tq4ch") pod "686edfbb-99ba-4e58-a789-ee673f2f4809" (UID: "686edfbb-99ba-4e58-a789-ee673f2f4809"). InnerVolumeSpecName "kube-api-access-tq4ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.116189 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-scripts" (OuterVolumeSpecName: "scripts") pod "686edfbb-99ba-4e58-a789-ee673f2f4809" (UID: "686edfbb-99ba-4e58-a789-ee673f2f4809"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.176083 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "686edfbb-99ba-4e58-a789-ee673f2f4809" (UID: "686edfbb-99ba-4e58-a789-ee673f2f4809"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.202350 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.202377 4848 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/686edfbb-99ba-4e58-a789-ee673f2f4809-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.202387 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq4ch\" (UniqueName: \"kubernetes.io/projected/686edfbb-99ba-4e58-a789-ee673f2f4809-kube-api-access-tq4ch\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.202400 4848 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.221742 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "686edfbb-99ba-4e58-a789-ee673f2f4809" (UID: "686edfbb-99ba-4e58-a789-ee673f2f4809"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.256091 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-config-data" (OuterVolumeSpecName: "config-data") pod "686edfbb-99ba-4e58-a789-ee673f2f4809" (UID: "686edfbb-99ba-4e58-a789-ee673f2f4809"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.304850 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.304895 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686edfbb-99ba-4e58-a789-ee673f2f4809-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.561119 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"686edfbb-99ba-4e58-a789-ee673f2f4809","Type":"ContainerDied","Data":"ffa91062ce86e11b50d5f95087bff6406ad36bc0923c9522663c49043f168ff7"} Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.561228 4848 scope.go:117] "RemoveContainer" containerID="f343cf2906706acd68e0505f577d5274cf8b0fd571cb0aa8908dc95474b2921a" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.561268 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.609221 4848 scope.go:117] "RemoveContainer" containerID="c5998e9cb9c10d8f31bb89d7415262142ac8d10b1b605c8b30129759efb9cedb" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.614836 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.641167 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.659657 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:11 crc kubenswrapper[4848]: E1204 14:14:11.660211 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerName="registry-server" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660235 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerName="registry-server" Dec 04 14:14:11 crc kubenswrapper[4848]: E1204 14:14:11.660260 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerName="extract-content" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660267 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerName="extract-content" Dec 04 14:14:11 crc kubenswrapper[4848]: E1204 14:14:11.660287 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="ceilometer-notification-agent" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660298 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="ceilometer-notification-agent" Dec 04 14:14:11 crc kubenswrapper[4848]: E1204 14:14:11.660337 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="ceilometer-central-agent" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660343 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="ceilometer-central-agent" Dec 04 14:14:11 crc kubenswrapper[4848]: E1204 14:14:11.660372 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="proxy-httpd" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660382 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="proxy-httpd" Dec 04 14:14:11 crc kubenswrapper[4848]: E1204 14:14:11.660404 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca5459c4-87fe-414e-9e6e-843d0e6329a9" containerName="heat-engine" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660416 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca5459c4-87fe-414e-9e6e-843d0e6329a9" containerName="heat-engine" Dec 04 14:14:11 crc kubenswrapper[4848]: E1204 14:14:11.660431 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerName="extract-utilities" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660438 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerName="extract-utilities" Dec 04 14:14:11 crc kubenswrapper[4848]: E1204 14:14:11.660457 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="sg-core" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660462 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="sg-core" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660782 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0f69f89-4463-4f8a-b892-ca81630f4b67" containerName="registry-server" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660806 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="ceilometer-central-agent" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660823 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="ceilometer-notification-agent" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660843 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca5459c4-87fe-414e-9e6e-843d0e6329a9" containerName="heat-engine" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660866 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="sg-core" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.660877 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" containerName="proxy-httpd" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.662870 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.670082 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.670287 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.687079 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.694012 4848 scope.go:117] "RemoveContainer" containerID="f3b4c16fc8876f0840e0e112fa1b0976398995341de26c5f47e141d66e2b620a" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.746179 4848 scope.go:117] "RemoveContainer" containerID="4df367335a2f8aa56f458612112434f462be68af1da23da029c11041c9d1ce39" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.817275 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb8rv\" (UniqueName: \"kubernetes.io/projected/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-kube-api-access-zb8rv\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.817338 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.817386 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-scripts\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.817410 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-run-httpd\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.817439 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-log-httpd\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.817472 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.817505 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-config-data\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.919096 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb8rv\" (UniqueName: \"kubernetes.io/projected/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-kube-api-access-zb8rv\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.919170 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.919211 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-scripts\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.919233 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-run-httpd\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.919255 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-log-httpd\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.919279 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.919306 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-config-data\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.920746 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-log-httpd\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.921410 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-run-httpd\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.927024 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.933555 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-scripts\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.937656 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.938552 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb8rv\" (UniqueName: \"kubernetes.io/projected/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-kube-api-access-zb8rv\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:11 crc kubenswrapper[4848]: I1204 14:14:11.940305 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-config-data\") pod \"ceilometer-0\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " pod="openstack/ceilometer-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.003992 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.144762 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.225988 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-config-data\") pod \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.226145 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk5bt\" (UniqueName: \"kubernetes.io/projected/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-kube-api-access-sk5bt\") pod \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.226235 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-scripts\") pod \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.226406 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-combined-ca-bundle\") pod \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\" (UID: \"013c9f72-2dcd-4e71-a4ab-aec59a8131aa\") " Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.232535 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-kube-api-access-sk5bt" (OuterVolumeSpecName: "kube-api-access-sk5bt") pod "013c9f72-2dcd-4e71-a4ab-aec59a8131aa" (UID: "013c9f72-2dcd-4e71-a4ab-aec59a8131aa"). InnerVolumeSpecName "kube-api-access-sk5bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.234192 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-scripts" (OuterVolumeSpecName: "scripts") pod "013c9f72-2dcd-4e71-a4ab-aec59a8131aa" (UID: "013c9f72-2dcd-4e71-a4ab-aec59a8131aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.274320 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-config-data" (OuterVolumeSpecName: "config-data") pod "013c9f72-2dcd-4e71-a4ab-aec59a8131aa" (UID: "013c9f72-2dcd-4e71-a4ab-aec59a8131aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.293731 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "013c9f72-2dcd-4e71-a4ab-aec59a8131aa" (UID: "013c9f72-2dcd-4e71-a4ab-aec59a8131aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.328772 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.328801 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.328811 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk5bt\" (UniqueName: \"kubernetes.io/projected/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-kube-api-access-sk5bt\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.328821 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/013c9f72-2dcd-4e71-a4ab-aec59a8131aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.409751 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="686edfbb-99ba-4e58-a789-ee673f2f4809" path="/var/lib/kubelet/pods/686edfbb-99ba-4e58-a789-ee673f2f4809/volumes" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.575045 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4czxk" event={"ID":"013c9f72-2dcd-4e71-a4ab-aec59a8131aa","Type":"ContainerDied","Data":"6327a5afb22da183a566e78ab5cd15137d2fc7609e40e888f97b6dc2e0845cab"} Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.575122 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6327a5afb22da183a566e78ab5cd15137d2fc7609e40e888f97b6dc2e0845cab" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.575199 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4czxk" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.638435 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.695726 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:12 crc kubenswrapper[4848]: E1204 14:14:12.696377 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013c9f72-2dcd-4e71-a4ab-aec59a8131aa" containerName="nova-cell0-conductor-db-sync" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.696396 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="013c9f72-2dcd-4e71-a4ab-aec59a8131aa" containerName="nova-cell0-conductor-db-sync" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.696609 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="013c9f72-2dcd-4e71-a4ab-aec59a8131aa" containerName="nova-cell0-conductor-db-sync" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.697426 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.700053 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.701377 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-48wpm" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.710552 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.839989 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp788\" (UniqueName: \"kubernetes.io/projected/143041c7-3ded-41b0-b60e-62cc2d1a61b0-kube-api-access-jp788\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.840292 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.840414 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.942644 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.942689 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.942774 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp788\" (UniqueName: \"kubernetes.io/projected/143041c7-3ded-41b0-b60e-62cc2d1a61b0-kube-api-access-jp788\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.949126 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.952677 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:12 crc kubenswrapper[4848]: I1204 14:14:12.966490 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp788\" (UniqueName: \"kubernetes.io/projected/143041c7-3ded-41b0-b60e-62cc2d1a61b0-kube-api-access-jp788\") pod \"nova-cell0-conductor-0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:13 crc kubenswrapper[4848]: I1204 14:14:13.036634 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:13 crc kubenswrapper[4848]: I1204 14:14:13.594468 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerStarted","Data":"f965aab7bfa3dfa757b082ceb59e1a5fe8bfea4d568bc71035d6915227078ef9"} Dec 04 14:14:13 crc kubenswrapper[4848]: I1204 14:14:13.594763 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerStarted","Data":"1aa0b0b91152c79719fe254a2264d63e4b0932bb18d6a50a96dc5f83bd227aae"} Dec 04 14:14:13 crc kubenswrapper[4848]: I1204 14:14:13.608671 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:14 crc kubenswrapper[4848]: I1204 14:14:14.314538 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:14:14 crc kubenswrapper[4848]: I1204 14:14:14.315336 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:14:14 crc kubenswrapper[4848]: I1204 14:14:14.409482 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:14 crc kubenswrapper[4848]: I1204 14:14:14.604889 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"143041c7-3ded-41b0-b60e-62cc2d1a61b0","Type":"ContainerStarted","Data":"f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df"} Dec 04 14:14:14 crc kubenswrapper[4848]: I1204 14:14:14.604930 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"143041c7-3ded-41b0-b60e-62cc2d1a61b0","Type":"ContainerStarted","Data":"191bb58107a530061abb19593ffcf9b886099f74753df6a45c593873e845b337"} Dec 04 14:14:14 crc kubenswrapper[4848]: I1204 14:14:14.606551 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:14 crc kubenswrapper[4848]: I1204 14:14:14.609275 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerStarted","Data":"0489b8d9e22b42a2e3ff908c5ae212509cea9893e90bd403df975611344b6769"} Dec 04 14:14:14 crc kubenswrapper[4848]: I1204 14:14:14.628056 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.628040093 podStartE2EDuration="2.628040093s" podCreationTimestamp="2025-12-04 14:14:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:14:14.618836009 +0000 UTC m=+1558.561332537" watchObservedRunningTime="2025-12-04 14:14:14.628040093 +0000 UTC m=+1558.570536621" Dec 04 14:14:15 crc kubenswrapper[4848]: I1204 14:14:15.624429 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerStarted","Data":"45ed6d19bf2e1a9a12d7e77aba7bb7314d4fb5be679d06d5ea780e836c6657b5"} Dec 04 14:14:15 crc kubenswrapper[4848]: I1204 14:14:15.789353 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:17 crc kubenswrapper[4848]: I1204 14:14:17.651443 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="143041c7-3ded-41b0-b60e-62cc2d1a61b0" containerName="nova-cell0-conductor-conductor" containerID="cri-o://f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df" gracePeriod=30 Dec 04 14:14:17 crc kubenswrapper[4848]: I1204 14:14:17.651901 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="ceilometer-central-agent" containerID="cri-o://f965aab7bfa3dfa757b082ceb59e1a5fe8bfea4d568bc71035d6915227078ef9" gracePeriod=30 Dec 04 14:14:17 crc kubenswrapper[4848]: I1204 14:14:17.652132 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="proxy-httpd" containerID="cri-o://9f6afa8b8c251f02e26fcf39ba0c93c256a5deb586e8cbeb90edd32b8e11dc68" gracePeriod=30 Dec 04 14:14:17 crc kubenswrapper[4848]: I1204 14:14:17.652190 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="sg-core" containerID="cri-o://45ed6d19bf2e1a9a12d7e77aba7bb7314d4fb5be679d06d5ea780e836c6657b5" gracePeriod=30 Dec 04 14:14:17 crc kubenswrapper[4848]: I1204 14:14:17.652224 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="ceilometer-notification-agent" containerID="cri-o://0489b8d9e22b42a2e3ff908c5ae212509cea9893e90bd403df975611344b6769" gracePeriod=30 Dec 04 14:14:17 crc kubenswrapper[4848]: I1204 14:14:17.652288 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerStarted","Data":"9f6afa8b8c251f02e26fcf39ba0c93c256a5deb586e8cbeb90edd32b8e11dc68"} Dec 04 14:14:17 crc kubenswrapper[4848]: I1204 14:14:17.683289 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.667545268 podStartE2EDuration="6.683270365s" podCreationTimestamp="2025-12-04 14:14:11 +0000 UTC" firstStartedPulling="2025-12-04 14:14:12.64018511 +0000 UTC m=+1556.582681638" lastFinishedPulling="2025-12-04 14:14:16.655910207 +0000 UTC m=+1560.598406735" observedRunningTime="2025-12-04 14:14:17.672229677 +0000 UTC m=+1561.614726195" watchObservedRunningTime="2025-12-04 14:14:17.683270365 +0000 UTC m=+1561.625766893" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.576739 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.673786 4848 generic.go:334] "Generic (PLEG): container finished" podID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerID="9f6afa8b8c251f02e26fcf39ba0c93c256a5deb586e8cbeb90edd32b8e11dc68" exitCode=0 Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.673831 4848 generic.go:334] "Generic (PLEG): container finished" podID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerID="45ed6d19bf2e1a9a12d7e77aba7bb7314d4fb5be679d06d5ea780e836c6657b5" exitCode=2 Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.673843 4848 generic.go:334] "Generic (PLEG): container finished" podID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerID="0489b8d9e22b42a2e3ff908c5ae212509cea9893e90bd403df975611344b6769" exitCode=0 Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.673889 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerDied","Data":"9f6afa8b8c251f02e26fcf39ba0c93c256a5deb586e8cbeb90edd32b8e11dc68"} Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.673923 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerDied","Data":"45ed6d19bf2e1a9a12d7e77aba7bb7314d4fb5be679d06d5ea780e836c6657b5"} Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.673938 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerDied","Data":"0489b8d9e22b42a2e3ff908c5ae212509cea9893e90bd403df975611344b6769"} Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.675246 4848 generic.go:334] "Generic (PLEG): container finished" podID="143041c7-3ded-41b0-b60e-62cc2d1a61b0" containerID="f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df" exitCode=0 Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.675280 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"143041c7-3ded-41b0-b60e-62cc2d1a61b0","Type":"ContainerDied","Data":"f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df"} Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.675302 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"143041c7-3ded-41b0-b60e-62cc2d1a61b0","Type":"ContainerDied","Data":"191bb58107a530061abb19593ffcf9b886099f74753df6a45c593873e845b337"} Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.675327 4848 scope.go:117] "RemoveContainer" containerID="f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.675487 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.702452 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp788\" (UniqueName: \"kubernetes.io/projected/143041c7-3ded-41b0-b60e-62cc2d1a61b0-kube-api-access-jp788\") pod \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.702537 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-config-data\") pod \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.702747 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-combined-ca-bundle\") pod \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\" (UID: \"143041c7-3ded-41b0-b60e-62cc2d1a61b0\") " Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.709146 4848 scope.go:117] "RemoveContainer" containerID="f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df" Dec 04 14:14:18 crc kubenswrapper[4848]: E1204 14:14:18.711348 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df\": container with ID starting with f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df not found: ID does not exist" containerID="f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.711393 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df"} err="failed to get container status \"f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df\": rpc error: code = NotFound desc = could not find container \"f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df\": container with ID starting with f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df not found: ID does not exist" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.711448 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143041c7-3ded-41b0-b60e-62cc2d1a61b0-kube-api-access-jp788" (OuterVolumeSpecName: "kube-api-access-jp788") pod "143041c7-3ded-41b0-b60e-62cc2d1a61b0" (UID: "143041c7-3ded-41b0-b60e-62cc2d1a61b0"). InnerVolumeSpecName "kube-api-access-jp788". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.745486 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "143041c7-3ded-41b0-b60e-62cc2d1a61b0" (UID: "143041c7-3ded-41b0-b60e-62cc2d1a61b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.748024 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-config-data" (OuterVolumeSpecName: "config-data") pod "143041c7-3ded-41b0-b60e-62cc2d1a61b0" (UID: "143041c7-3ded-41b0-b60e-62cc2d1a61b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.806053 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.806101 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp788\" (UniqueName: \"kubernetes.io/projected/143041c7-3ded-41b0-b60e-62cc2d1a61b0-kube-api-access-jp788\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:18 crc kubenswrapper[4848]: I1204 14:14:18.806117 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/143041c7-3ded-41b0-b60e-62cc2d1a61b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.006854 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.017585 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.069010 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:19 crc kubenswrapper[4848]: E1204 14:14:19.069521 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143041c7-3ded-41b0-b60e-62cc2d1a61b0" containerName="nova-cell0-conductor-conductor" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.069542 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="143041c7-3ded-41b0-b60e-62cc2d1a61b0" containerName="nova-cell0-conductor-conductor" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.069795 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="143041c7-3ded-41b0-b60e-62cc2d1a61b0" containerName="nova-cell0-conductor-conductor" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.070571 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.073589 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-48wpm" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.073742 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.082694 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.214446 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae44733-c46b-4ed7-9371-377d026bc216-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.214504 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2ptv\" (UniqueName: \"kubernetes.io/projected/7ae44733-c46b-4ed7-9371-377d026bc216-kube-api-access-z2ptv\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.214610 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae44733-c46b-4ed7-9371-377d026bc216-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.316457 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae44733-c46b-4ed7-9371-377d026bc216-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.316759 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae44733-c46b-4ed7-9371-377d026bc216-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.316812 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2ptv\" (UniqueName: \"kubernetes.io/projected/7ae44733-c46b-4ed7-9371-377d026bc216-kube-api-access-z2ptv\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.321200 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae44733-c46b-4ed7-9371-377d026bc216-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.326580 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae44733-c46b-4ed7-9371-377d026bc216-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.335590 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2ptv\" (UniqueName: \"kubernetes.io/projected/7ae44733-c46b-4ed7-9371-377d026bc216-kube-api-access-z2ptv\") pod \"nova-cell0-conductor-0\" (UID: \"7ae44733-c46b-4ed7-9371-377d026bc216\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.392565 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:19 crc kubenswrapper[4848]: I1204 14:14:19.968711 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:14:20 crc kubenswrapper[4848]: I1204 14:14:20.405780 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="143041c7-3ded-41b0-b60e-62cc2d1a61b0" path="/var/lib/kubelet/pods/143041c7-3ded-41b0-b60e-62cc2d1a61b0/volumes" Dec 04 14:14:20 crc kubenswrapper[4848]: I1204 14:14:20.746223 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7ae44733-c46b-4ed7-9371-377d026bc216","Type":"ContainerStarted","Data":"01b6276bfa63f3259c7b71b6db820305b140571565c8ddb794442760d1c3bde4"} Dec 04 14:14:20 crc kubenswrapper[4848]: I1204 14:14:20.746277 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7ae44733-c46b-4ed7-9371-377d026bc216","Type":"ContainerStarted","Data":"b30e90b081b8728de34eedbb2a027e2b48e19aaaefb7c7c5e236c4889a0d8437"} Dec 04 14:14:20 crc kubenswrapper[4848]: I1204 14:14:20.746912 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:20 crc kubenswrapper[4848]: I1204 14:14:20.775775 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.775756221 podStartE2EDuration="1.775756221s" podCreationTimestamp="2025-12-04 14:14:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:14:20.762882179 +0000 UTC m=+1564.705378707" watchObservedRunningTime="2025-12-04 14:14:20.775756221 +0000 UTC m=+1564.718252749" Dec 04 14:14:23 crc kubenswrapper[4848]: I1204 14:14:23.799150 4848 generic.go:334] "Generic (PLEG): container finished" podID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerID="f965aab7bfa3dfa757b082ceb59e1a5fe8bfea4d568bc71035d6915227078ef9" exitCode=0 Dec 04 14:14:23 crc kubenswrapper[4848]: I1204 14:14:23.799680 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerDied","Data":"f965aab7bfa3dfa757b082ceb59e1a5fe8bfea4d568bc71035d6915227078ef9"} Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.012534 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.143038 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-combined-ca-bundle\") pod \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.143122 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-log-httpd\") pod \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.143192 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-scripts\") pod \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.143249 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-run-httpd\") pod \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.143292 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb8rv\" (UniqueName: \"kubernetes.io/projected/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-kube-api-access-zb8rv\") pod \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.143329 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-config-data\") pod \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.143343 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-sg-core-conf-yaml\") pod \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\" (UID: \"bc3a7415-22f5-4d0f-8939-00a7427a6d2f\") " Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.143787 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bc3a7415-22f5-4d0f-8939-00a7427a6d2f" (UID: "bc3a7415-22f5-4d0f-8939-00a7427a6d2f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.144194 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bc3a7415-22f5-4d0f-8939-00a7427a6d2f" (UID: "bc3a7415-22f5-4d0f-8939-00a7427a6d2f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.145165 4848 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.145198 4848 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.163625 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-scripts" (OuterVolumeSpecName: "scripts") pod "bc3a7415-22f5-4d0f-8939-00a7427a6d2f" (UID: "bc3a7415-22f5-4d0f-8939-00a7427a6d2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.163695 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-kube-api-access-zb8rv" (OuterVolumeSpecName: "kube-api-access-zb8rv") pod "bc3a7415-22f5-4d0f-8939-00a7427a6d2f" (UID: "bc3a7415-22f5-4d0f-8939-00a7427a6d2f"). InnerVolumeSpecName "kube-api-access-zb8rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.183750 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bc3a7415-22f5-4d0f-8939-00a7427a6d2f" (UID: "bc3a7415-22f5-4d0f-8939-00a7427a6d2f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.249285 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.249568 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb8rv\" (UniqueName: \"kubernetes.io/projected/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-kube-api-access-zb8rv\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.249666 4848 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.270112 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc3a7415-22f5-4d0f-8939-00a7427a6d2f" (UID: "bc3a7415-22f5-4d0f-8939-00a7427a6d2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.304641 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-config-data" (OuterVolumeSpecName: "config-data") pod "bc3a7415-22f5-4d0f-8939-00a7427a6d2f" (UID: "bc3a7415-22f5-4d0f-8939-00a7427a6d2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.351884 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.352216 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3a7415-22f5-4d0f-8939-00a7427a6d2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.818367 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc3a7415-22f5-4d0f-8939-00a7427a6d2f","Type":"ContainerDied","Data":"1aa0b0b91152c79719fe254a2264d63e4b0932bb18d6a50a96dc5f83bd227aae"} Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.818425 4848 scope.go:117] "RemoveContainer" containerID="9f6afa8b8c251f02e26fcf39ba0c93c256a5deb586e8cbeb90edd32b8e11dc68" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.818589 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.855249 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.863286 4848 scope.go:117] "RemoveContainer" containerID="45ed6d19bf2e1a9a12d7e77aba7bb7314d4fb5be679d06d5ea780e836c6657b5" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.873530 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.890336 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.890503 4848 scope.go:117] "RemoveContainer" containerID="0489b8d9e22b42a2e3ff908c5ae212509cea9893e90bd403df975611344b6769" Dec 04 14:14:24 crc kubenswrapper[4848]: E1204 14:14:24.890914 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="proxy-httpd" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.890970 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="proxy-httpd" Dec 04 14:14:24 crc kubenswrapper[4848]: E1204 14:14:24.890999 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="sg-core" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.891008 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="sg-core" Dec 04 14:14:24 crc kubenswrapper[4848]: E1204 14:14:24.891040 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="ceilometer-central-agent" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.891049 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="ceilometer-central-agent" Dec 04 14:14:24 crc kubenswrapper[4848]: E1204 14:14:24.891067 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="ceilometer-notification-agent" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.891075 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="ceilometer-notification-agent" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.891349 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="ceilometer-notification-agent" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.891372 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="proxy-httpd" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.891388 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="sg-core" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.891399 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" containerName="ceilometer-central-agent" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.893454 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.900095 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.902680 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.902769 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:14:24 crc kubenswrapper[4848]: I1204 14:14:24.930352 4848 scope.go:117] "RemoveContainer" containerID="f965aab7bfa3dfa757b082ceb59e1a5fe8bfea4d568bc71035d6915227078ef9" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.067142 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-run-httpd\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.067180 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-log-httpd\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.067606 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-config-data\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.067741 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdhzg\" (UniqueName: \"kubernetes.io/projected/20711d1f-8dec-41e1-920e-50603c2c7b79-kube-api-access-zdhzg\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.068111 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.068162 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-scripts\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.068233 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.170170 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-config-data\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.170232 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdhzg\" (UniqueName: \"kubernetes.io/projected/20711d1f-8dec-41e1-920e-50603c2c7b79-kube-api-access-zdhzg\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.170674 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.170701 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-scripts\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.171278 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.171372 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-run-httpd\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.171412 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-log-httpd\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.171972 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-log-httpd\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.171975 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-run-httpd\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.174365 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.174505 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.175065 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-scripts\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.178268 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-config-data\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.200340 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdhzg\" (UniqueName: \"kubernetes.io/projected/20711d1f-8dec-41e1-920e-50603c2c7b79-kube-api-access-zdhzg\") pod \"ceilometer-0\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.217997 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.734676 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:14:25 crc kubenswrapper[4848]: I1204 14:14:25.829984 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerStarted","Data":"1144c708ff4dde3da8e4a63694c8705103a7c845a8a4b6c9e4bea32d229d1d61"} Dec 04 14:14:26 crc kubenswrapper[4848]: I1204 14:14:26.405504 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc3a7415-22f5-4d0f-8939-00a7427a6d2f" path="/var/lib/kubelet/pods/bc3a7415-22f5-4d0f-8939-00a7427a6d2f/volumes" Dec 04 14:14:26 crc kubenswrapper[4848]: I1204 14:14:26.844266 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerStarted","Data":"77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c"} Dec 04 14:14:27 crc kubenswrapper[4848]: I1204 14:14:27.856617 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerStarted","Data":"ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e"} Dec 04 14:14:28 crc kubenswrapper[4848]: I1204 14:14:28.886195 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerStarted","Data":"b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495"} Dec 04 14:14:29 crc kubenswrapper[4848]: I1204 14:14:29.444818 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 14:14:29 crc kubenswrapper[4848]: I1204 14:14:29.900059 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerStarted","Data":"4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160"} Dec 04 14:14:29 crc kubenswrapper[4848]: I1204 14:14:29.900317 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.137801 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.067856295 podStartE2EDuration="6.137781883s" podCreationTimestamp="2025-12-04 14:14:24 +0000 UTC" firstStartedPulling="2025-12-04 14:14:25.731066496 +0000 UTC m=+1569.673563024" lastFinishedPulling="2025-12-04 14:14:28.800992084 +0000 UTC m=+1572.743488612" observedRunningTime="2025-12-04 14:14:29.959235139 +0000 UTC m=+1573.901731657" watchObservedRunningTime="2025-12-04 14:14:30.137781883 +0000 UTC m=+1574.080278411" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.138785 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-s9gd9"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.140189 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.141728 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.142708 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.165181 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-s9gd9"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.312531 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.312587 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zznk2\" (UniqueName: \"kubernetes.io/projected/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-kube-api-access-zznk2\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.312650 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-scripts\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.312728 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-config-data\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.355656 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.362126 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.367326 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.383007 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.414372 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-scripts\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.414486 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-config-data\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.414579 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.414612 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zznk2\" (UniqueName: \"kubernetes.io/projected/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-kube-api-access-zznk2\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.418713 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.433847 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.437397 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.439513 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-scripts\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.451516 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-config-data\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.464841 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.465389 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-sl25v"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.466823 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.476578 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zznk2\" (UniqueName: \"kubernetes.io/projected/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-kube-api-access-zznk2\") pod \"nova-cell0-cell-mapping-s9gd9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.511026 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.511531 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.527660 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-logs\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.527989 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-config-data\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.528128 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.528274 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtgh9\" (UniqueName: \"kubernetes.io/projected/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-kube-api-access-rtgh9\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.610434 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-sl25v"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.631967 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-config-data\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.632032 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjh89\" (UniqueName: \"kubernetes.io/projected/30e84076-e3b2-4bac-b59a-9085a8f256ff-kube-api-access-sjh89\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.632091 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92mrb\" (UniqueName: \"kubernetes.io/projected/db1d9746-4bfb-48fb-b965-bf89eb95885c-kube-api-access-92mrb\") pod \"aodh-db-create-sl25v\" (UID: \"db1d9746-4bfb-48fb-b965-bf89eb95885c\") " pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.632210 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.632374 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtgh9\" (UniqueName: \"kubernetes.io/projected/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-kube-api-access-rtgh9\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.632451 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-logs\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.632565 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.632597 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-config-data\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.632790 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db1d9746-4bfb-48fb-b965-bf89eb95885c-operator-scripts\") pod \"aodh-db-create-sl25v\" (UID: \"db1d9746-4bfb-48fb-b965-bf89eb95885c\") " pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.635221 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-logs\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.635291 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.637433 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.662748 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-config-data\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.666362 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.685094 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.691778 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtgh9\" (UniqueName: \"kubernetes.io/projected/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-kube-api-access-rtgh9\") pod \"nova-api-0\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.708245 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.752886 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92mrb\" (UniqueName: \"kubernetes.io/projected/db1d9746-4bfb-48fb-b965-bf89eb95885c-kube-api-access-92mrb\") pod \"aodh-db-create-sl25v\" (UID: \"db1d9746-4bfb-48fb-b965-bf89eb95885c\") " pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.752995 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8583482-412b-4ab5-8089-6eb3587cc0ef-logs\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.753116 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.753141 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-config-data\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.753174 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-config-data\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.753199 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.753229 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zltlz\" (UniqueName: \"kubernetes.io/projected/c8583482-412b-4ab5-8089-6eb3587cc0ef-kube-api-access-zltlz\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.753283 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db1d9746-4bfb-48fb-b965-bf89eb95885c-operator-scripts\") pod \"aodh-db-create-sl25v\" (UID: \"db1d9746-4bfb-48fb-b965-bf89eb95885c\") " pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.753311 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjh89\" (UniqueName: \"kubernetes.io/projected/30e84076-e3b2-4bac-b59a-9085a8f256ff-kube-api-access-sjh89\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.758475 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.759170 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db1d9746-4bfb-48fb-b965-bf89eb95885c-operator-scripts\") pod \"aodh-db-create-sl25v\" (UID: \"db1d9746-4bfb-48fb-b965-bf89eb95885c\") " pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.760871 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.764131 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-config-data\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.788174 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92mrb\" (UniqueName: \"kubernetes.io/projected/db1d9746-4bfb-48fb-b965-bf89eb95885c-kube-api-access-92mrb\") pod \"aodh-db-create-sl25v\" (UID: \"db1d9746-4bfb-48fb-b965-bf89eb95885c\") " pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.790519 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjh89\" (UniqueName: \"kubernetes.io/projected/30e84076-e3b2-4bac-b59a-9085a8f256ff-kube-api-access-sjh89\") pod \"nova-scheduler-0\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.801254 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.802666 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.808195 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.822993 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-fb6e-account-create-update-sdrw9"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.824547 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.829322 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.843664 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855418 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-config-data\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855445 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855470 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zltlz\" (UniqueName: \"kubernetes.io/projected/c8583482-412b-4ab5-8089-6eb3587cc0ef-kube-api-access-zltlz\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855494 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855584 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855620 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8583482-412b-4ab5-8089-6eb3587cc0ef-logs\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855636 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g5hb\" (UniqueName: \"kubernetes.io/projected/5fff65b6-de45-4e47-9224-d8620d7ae78e-kube-api-access-6g5hb\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855657 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhts5\" (UniqueName: \"kubernetes.io/projected/d6b33655-7b9c-4663-b8d0-85be8361f65f-kube-api-access-zhts5\") pod \"aodh-fb6e-account-create-update-sdrw9\" (UID: \"d6b33655-7b9c-4663-b8d0-85be8361f65f\") " pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.855703 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6b33655-7b9c-4663-b8d0-85be8361f65f-operator-scripts\") pod \"aodh-fb6e-account-create-update-sdrw9\" (UID: \"d6b33655-7b9c-4663-b8d0-85be8361f65f\") " pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.862363 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8583482-412b-4ab5-8089-6eb3587cc0ef-logs\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.863562 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-config-data\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.868803 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.877514 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-fb6e-account-create-update-sdrw9"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.893098 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zltlz\" (UniqueName: \"kubernetes.io/projected/c8583482-412b-4ab5-8089-6eb3587cc0ef-kube-api-access-zltlz\") pod \"nova-metadata-0\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " pod="openstack/nova-metadata-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.976759 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6b33655-7b9c-4663-b8d0-85be8361f65f-operator-scripts\") pod \"aodh-fb6e-account-create-update-sdrw9\" (UID: \"d6b33655-7b9c-4663-b8d0-85be8361f65f\") " pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.979030 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.979200 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.979267 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g5hb\" (UniqueName: \"kubernetes.io/projected/5fff65b6-de45-4e47-9224-d8620d7ae78e-kube-api-access-6g5hb\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.979288 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhts5\" (UniqueName: \"kubernetes.io/projected/d6b33655-7b9c-4663-b8d0-85be8361f65f-kube-api-access-zhts5\") pod \"aodh-fb6e-account-create-update-sdrw9\" (UID: \"d6b33655-7b9c-4663-b8d0-85be8361f65f\") " pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.980106 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6b33655-7b9c-4663-b8d0-85be8361f65f-operator-scripts\") pod \"aodh-fb6e-account-create-update-sdrw9\" (UID: \"d6b33655-7b9c-4663-b8d0-85be8361f65f\") " pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.990908 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-rx28f"] Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.994451 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:30 crc kubenswrapper[4848]: I1204 14:14:30.995328 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.008118 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.017572 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhts5\" (UniqueName: \"kubernetes.io/projected/d6b33655-7b9c-4663-b8d0-85be8361f65f-kube-api-access-zhts5\") pod \"aodh-fb6e-account-create-update-sdrw9\" (UID: \"d6b33655-7b9c-4663-b8d0-85be8361f65f\") " pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.027991 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.036272 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g5hb\" (UniqueName: \"kubernetes.io/projected/5fff65b6-de45-4e47-9224-d8620d7ae78e-kube-api-access-6g5hb\") pod \"nova-cell1-novncproxy-0\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.062319 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-rx28f"] Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.072853 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.084218 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzgm8\" (UniqueName: \"kubernetes.io/projected/576a9a23-e104-4d55-8100-ee23628af7cf-kube-api-access-pzgm8\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.084375 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-config\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.084409 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.084565 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-svc\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.084981 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.085026 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.089517 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.134452 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.152633 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.187641 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-svc\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.187778 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.187809 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.187967 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzgm8\" (UniqueName: \"kubernetes.io/projected/576a9a23-e104-4d55-8100-ee23628af7cf-kube-api-access-pzgm8\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.188051 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-config\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.188083 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.189536 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.189630 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.190792 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-config\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.190898 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-svc\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.191085 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.217707 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzgm8\" (UniqueName: \"kubernetes.io/projected/576a9a23-e104-4d55-8100-ee23628af7cf-kube-api-access-pzgm8\") pod \"dnsmasq-dns-9b86998b5-rx28f\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.341990 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.574038 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8tc6k"] Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.575748 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.578941 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.579193 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.600061 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8tc6k"] Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.601781 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkzhh\" (UniqueName: \"kubernetes.io/projected/cb391cd3-6f78-4e51-ac06-14c9420d0f06-kube-api-access-dkzhh\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.601860 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-scripts\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.601895 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-config-data\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.601915 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.647400 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-s9gd9"] Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.661797 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.704049 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-config-data\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.704361 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.705022 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkzhh\" (UniqueName: \"kubernetes.io/projected/cb391cd3-6f78-4e51-ac06-14c9420d0f06-kube-api-access-dkzhh\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.705132 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-scripts\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.722900 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.739799 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-config-data\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.756254 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-scripts\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.756902 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkzhh\" (UniqueName: \"kubernetes.io/projected/cb391cd3-6f78-4e51-ac06-14c9420d0f06-kube-api-access-dkzhh\") pod \"nova-cell1-conductor-db-sync-8tc6k\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.789500 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:31 crc kubenswrapper[4848]: I1204 14:14:31.855578 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:32 crc kubenswrapper[4848]: I1204 14:14:32.024139 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf","Type":"ContainerStarted","Data":"0bfef95fba0da49d2a8133742527d7fe09eae63b39e16394092c9693052700e5"} Dec 04 14:14:32 crc kubenswrapper[4848]: I1204 14:14:32.054719 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-s9gd9" event={"ID":"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9","Type":"ContainerStarted","Data":"ba9a09732f25d902e3edd21a5826b136f5bd2e935008728575ec93de93984765"} Dec 04 14:14:32 crc kubenswrapper[4848]: I1204 14:14:32.090932 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"30e84076-e3b2-4bac-b59a-9085a8f256ff","Type":"ContainerStarted","Data":"6080cd82eedbcc3ade843ffa1c01d365b8ee74c279f7c3a66b00d0b99a8e40df"} Dec 04 14:14:32 crc kubenswrapper[4848]: I1204 14:14:32.115475 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-sl25v"] Dec 04 14:14:32 crc kubenswrapper[4848]: I1204 14:14:32.619111 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-rx28f"] Dec 04 14:14:32 crc kubenswrapper[4848]: I1204 14:14:32.723924 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:14:32 crc kubenswrapper[4848]: W1204 14:14:32.757576 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice/crio-9f6afa8b8c251f02e26fcf39ba0c93c256a5deb586e8cbeb90edd32b8e11dc68.scope WatchSource:0}: Error finding container 9f6afa8b8c251f02e26fcf39ba0c93c256a5deb586e8cbeb90edd32b8e11dc68: Status 404 returned error can't find the container with id 9f6afa8b8c251f02e26fcf39ba0c93c256a5deb586e8cbeb90edd32b8e11dc68 Dec 04 14:14:32 crc kubenswrapper[4848]: I1204 14:14:32.770561 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-fb6e-account-create-update-sdrw9"] Dec 04 14:14:32 crc kubenswrapper[4848]: I1204 14:14:32.800097 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.089793 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8tc6k"] Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.154260 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8583482-412b-4ab5-8089-6eb3587cc0ef","Type":"ContainerStarted","Data":"4a2126b2c5abeef91f710cc4275d10ba611f571bb15951a61481014409ef1daf"} Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.156993 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fff65b6-de45-4e47-9224-d8620d7ae78e","Type":"ContainerStarted","Data":"f9ac3e779da587ed1e411554738d5b21a544bab4f7d0c041205749ec343677f4"} Dec 04 14:14:33 crc kubenswrapper[4848]: W1204 14:14:33.192085 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb391cd3_6f78_4e51_ac06_14c9420d0f06.slice/crio-5c9ac4b60c8d9897d70e966265a3a59e1e0f2d0222e65a90b351ccbbb6761763 WatchSource:0}: Error finding container 5c9ac4b60c8d9897d70e966265a3a59e1e0f2d0222e65a90b351ccbbb6761763: Status 404 returned error can't find the container with id 5c9ac4b60c8d9897d70e966265a3a59e1e0f2d0222e65a90b351ccbbb6761763 Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.192381 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-s9gd9" event={"ID":"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9","Type":"ContainerStarted","Data":"e6bb0c6f78150ddbeb78da3883863401a04086ec3f231a8cdcfeaf559006f080"} Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.203018 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" event={"ID":"576a9a23-e104-4d55-8100-ee23628af7cf","Type":"ContainerStarted","Data":"547e7e3cd3211e334c27881428a32645df5ac620fe0df7250b0de191e7bfe063"} Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.233128 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-s9gd9" podStartSLOduration=3.233108199 podStartE2EDuration="3.233108199s" podCreationTimestamp="2025-12-04 14:14:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:14:33.213560524 +0000 UTC m=+1577.156057052" watchObservedRunningTime="2025-12-04 14:14:33.233108199 +0000 UTC m=+1577.175604727" Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.245750 4848 generic.go:334] "Generic (PLEG): container finished" podID="db1d9746-4bfb-48fb-b965-bf89eb95885c" containerID="ebc3b009cf446e78b02aa4807272e0de9c1e2ac6667948d1e45f6b12e02c568b" exitCode=0 Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.245834 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-sl25v" event={"ID":"db1d9746-4bfb-48fb-b965-bf89eb95885c","Type":"ContainerDied","Data":"ebc3b009cf446e78b02aa4807272e0de9c1e2ac6667948d1e45f6b12e02c568b"} Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.245866 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-sl25v" event={"ID":"db1d9746-4bfb-48fb-b965-bf89eb95885c","Type":"ContainerStarted","Data":"11f5f425f2d192fb11599370b0102114e94f5503507185c84005900872327141"} Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.273152 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-fb6e-account-create-update-sdrw9" event={"ID":"d6b33655-7b9c-4663-b8d0-85be8361f65f","Type":"ContainerStarted","Data":"759b3b4aaa784769502708de23fad28653803e95715dcae96dd9dc3ef86357d1"} Dec 04 14:14:33 crc kubenswrapper[4848]: E1204 14:14:33.275869 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod143041c7_3ded_41b0_b60e_62cc2d1a61b0.slice/crio-191bb58107a530061abb19593ffcf9b886099f74753df6a45c593873e845b337\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70bfcde6_2ad2_44fc_ae39_5821c63f4918.slice/crio-4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice/crio-conmon-f965aab7bfa3dfa757b082ceb59e1a5fe8bfea4d568bc71035d6915227078ef9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod143041c7_3ded_41b0_b60e_62cc2d1a61b0.slice/crio-conmon-f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice/crio-0489b8d9e22b42a2e3ff908c5ae212509cea9893e90bd403df975611344b6769.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb38b0b7c_d339_418d_b2c4_0278a451626a.slice/crio-27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb38b0b7c_d339_418d_b2c4_0278a451626a.slice/crio-conmon-27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice/crio-conmon-45ed6d19bf2e1a9a12d7e77aba7bb7314d4fb5be679d06d5ea780e836c6657b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70bfcde6_2ad2_44fc_ae39_5821c63f4918.slice/crio-conmon-4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice/crio-conmon-0489b8d9e22b42a2e3ff908c5ae212509cea9893e90bd403df975611344b6769.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod143041c7_3ded_41b0_b60e_62cc2d1a61b0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice/crio-f965aab7bfa3dfa757b082ceb59e1a5fe8bfea4d568bc71035d6915227078ef9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice/crio-1aa0b0b91152c79719fe254a2264d63e4b0932bb18d6a50a96dc5f83bd227aae\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod143041c7_3ded_41b0_b60e_62cc2d1a61b0.slice/crio-f71a1f2a74301ced82f1767d2b2eb8e589975b5779673cec87dc7dcf558f94df.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3a7415_22f5_4d0f_8939_00a7427a6d2f.slice/crio-45ed6d19bf2e1a9a12d7e77aba7bb7314d4fb5be679d06d5ea780e836c6657b5.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:14:33 crc kubenswrapper[4848]: E1204 14:14:33.284754 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb38b0b7c_d339_418d_b2c4_0278a451626a.slice/crio-27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb38b0b7c_d339_418d_b2c4_0278a451626a.slice/crio-conmon-27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb1d9746_4bfb_48fb_b965_bf89eb95885c.slice/crio-ebc3b009cf446e78b02aa4807272e0de9c1e2ac6667948d1e45f6b12e02c568b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70bfcde6_2ad2_44fc_ae39_5821c63f4918.slice/crio-conmon-4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70bfcde6_2ad2_44fc_ae39_5821c63f4918.slice/crio-4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.309342 4848 generic.go:334] "Generic (PLEG): container finished" podID="70bfcde6-2ad2-44fc-ae39-5821c63f4918" containerID="4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc" exitCode=137 Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.309441 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-65b648c688-s9c2l" event={"ID":"70bfcde6-2ad2-44fc-ae39-5821c63f4918","Type":"ContainerDied","Data":"4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc"} Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.358246 4848 generic.go:334] "Generic (PLEG): container finished" podID="b38b0b7c-d339-418d-b2c4-0278a451626a" containerID="27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa" exitCode=137 Dec 04 14:14:33 crc kubenswrapper[4848]: I1204 14:14:33.358429 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" event={"ID":"b38b0b7c-d339-418d-b2c4-0278a451626a","Type":"ContainerDied","Data":"27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa"} Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.010653 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.039937 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.130503 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data\") pod \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.130634 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data\") pod \"b38b0b7c-d339-418d-b2c4-0278a451626a\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.130758 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data-custom\") pod \"b38b0b7c-d339-418d-b2c4-0278a451626a\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.130866 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpvbv\" (UniqueName: \"kubernetes.io/projected/b38b0b7c-d339-418d-b2c4-0278a451626a-kube-api-access-cpvbv\") pod \"b38b0b7c-d339-418d-b2c4-0278a451626a\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.130916 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-combined-ca-bundle\") pod \"b38b0b7c-d339-418d-b2c4-0278a451626a\" (UID: \"b38b0b7c-d339-418d-b2c4-0278a451626a\") " Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.131070 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-combined-ca-bundle\") pod \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.131144 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data-custom\") pod \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.131233 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kwpz\" (UniqueName: \"kubernetes.io/projected/70bfcde6-2ad2-44fc-ae39-5821c63f4918-kube-api-access-4kwpz\") pod \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\" (UID: \"70bfcde6-2ad2-44fc-ae39-5821c63f4918\") " Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.140673 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "70bfcde6-2ad2-44fc-ae39-5821c63f4918" (UID: "70bfcde6-2ad2-44fc-ae39-5821c63f4918"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.157417 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70bfcde6-2ad2-44fc-ae39-5821c63f4918-kube-api-access-4kwpz" (OuterVolumeSpecName: "kube-api-access-4kwpz") pod "70bfcde6-2ad2-44fc-ae39-5821c63f4918" (UID: "70bfcde6-2ad2-44fc-ae39-5821c63f4918"). InnerVolumeSpecName "kube-api-access-4kwpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.157485 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b38b0b7c-d339-418d-b2c4-0278a451626a-kube-api-access-cpvbv" (OuterVolumeSpecName: "kube-api-access-cpvbv") pod "b38b0b7c-d339-418d-b2c4-0278a451626a" (UID: "b38b0b7c-d339-418d-b2c4-0278a451626a"). InnerVolumeSpecName "kube-api-access-cpvbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.178626 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b38b0b7c-d339-418d-b2c4-0278a451626a" (UID: "b38b0b7c-d339-418d-b2c4-0278a451626a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.197155 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70bfcde6-2ad2-44fc-ae39-5821c63f4918" (UID: "70bfcde6-2ad2-44fc-ae39-5821c63f4918"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.230130 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b38b0b7c-d339-418d-b2c4-0278a451626a" (UID: "b38b0b7c-d339-418d-b2c4-0278a451626a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.239303 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpvbv\" (UniqueName: \"kubernetes.io/projected/b38b0b7c-d339-418d-b2c4-0278a451626a-kube-api-access-cpvbv\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.239347 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.239361 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.239374 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.239386 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kwpz\" (UniqueName: \"kubernetes.io/projected/70bfcde6-2ad2-44fc-ae39-5821c63f4918-kube-api-access-4kwpz\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.239401 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.267044 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data" (OuterVolumeSpecName: "config-data") pod "70bfcde6-2ad2-44fc-ae39-5821c63f4918" (UID: "70bfcde6-2ad2-44fc-ae39-5821c63f4918"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.292847 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data" (OuterVolumeSpecName: "config-data") pod "b38b0b7c-d339-418d-b2c4-0278a451626a" (UID: "b38b0b7c-d339-418d-b2c4-0278a451626a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.344552 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bfcde6-2ad2-44fc-ae39-5821c63f4918-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.344587 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b38b0b7c-d339-418d-b2c4-0278a451626a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.377757 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" event={"ID":"cb391cd3-6f78-4e51-ac06-14c9420d0f06","Type":"ContainerStarted","Data":"efc73cb5821d728230b71deab4febbbdd8471088fbe6454ec39fa4a2f68e23ac"} Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.377802 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" event={"ID":"cb391cd3-6f78-4e51-ac06-14c9420d0f06","Type":"ContainerStarted","Data":"5c9ac4b60c8d9897d70e966265a3a59e1e0f2d0222e65a90b351ccbbb6761763"} Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.386915 4848 generic.go:334] "Generic (PLEG): container finished" podID="d6b33655-7b9c-4663-b8d0-85be8361f65f" containerID="68bd1d83369d6855828689b87806360391bad180ef3c9372f79e96228be122f4" exitCode=0 Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.387017 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-fb6e-account-create-update-sdrw9" event={"ID":"d6b33655-7b9c-4663-b8d0-85be8361f65f","Type":"ContainerDied","Data":"68bd1d83369d6855828689b87806360391bad180ef3c9372f79e96228be122f4"} Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.391049 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-65b648c688-s9c2l" event={"ID":"70bfcde6-2ad2-44fc-ae39-5821c63f4918","Type":"ContainerDied","Data":"4df30f9ebbee66082ff0a6f20c3e1bcfa1b58470025e84b96a89d379b6db8cd0"} Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.391095 4848 scope.go:117] "RemoveContainer" containerID="4411c5fb14ff7e49ffa5b7a2b7e4eac9b91e2d9d7815604e1dd4c7f0e6352ebc" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.391217 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-65b648c688-s9c2l" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.405318 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" podStartSLOduration=3.405301192 podStartE2EDuration="3.405301192s" podCreationTimestamp="2025-12-04 14:14:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:14:34.402142116 +0000 UTC m=+1578.344638654" watchObservedRunningTime="2025-12-04 14:14:34.405301192 +0000 UTC m=+1578.347797720" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.406310 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.415024 4848 generic.go:334] "Generic (PLEG): container finished" podID="576a9a23-e104-4d55-8100-ee23628af7cf" containerID="e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7" exitCode=0 Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.468041 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-665fbf8d68-zvxk9" event={"ID":"b38b0b7c-d339-418d-b2c4-0278a451626a","Type":"ContainerDied","Data":"534c15b09d5ca3658c48975811ff39749c6fd3c04024da2536f2f57e74ea48c9"} Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.468091 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-65b648c688-s9c2l"] Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.468108 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" event={"ID":"576a9a23-e104-4d55-8100-ee23628af7cf","Type":"ContainerDied","Data":"e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7"} Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.500303 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-65b648c688-s9c2l"] Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.544705 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-665fbf8d68-zvxk9"] Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.558941 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-665fbf8d68-zvxk9"] Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.857678 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:14:34 crc kubenswrapper[4848]: I1204 14:14:34.874909 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:14:35 crc kubenswrapper[4848]: I1204 14:14:35.662454 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:35 crc kubenswrapper[4848]: I1204 14:14:35.801161 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92mrb\" (UniqueName: \"kubernetes.io/projected/db1d9746-4bfb-48fb-b965-bf89eb95885c-kube-api-access-92mrb\") pod \"db1d9746-4bfb-48fb-b965-bf89eb95885c\" (UID: \"db1d9746-4bfb-48fb-b965-bf89eb95885c\") " Dec 04 14:14:35 crc kubenswrapper[4848]: I1204 14:14:35.801404 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db1d9746-4bfb-48fb-b965-bf89eb95885c-operator-scripts\") pod \"db1d9746-4bfb-48fb-b965-bf89eb95885c\" (UID: \"db1d9746-4bfb-48fb-b965-bf89eb95885c\") " Dec 04 14:14:35 crc kubenswrapper[4848]: I1204 14:14:35.802274 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db1d9746-4bfb-48fb-b965-bf89eb95885c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db1d9746-4bfb-48fb-b965-bf89eb95885c" (UID: "db1d9746-4bfb-48fb-b965-bf89eb95885c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:14:35 crc kubenswrapper[4848]: I1204 14:14:35.808681 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db1d9746-4bfb-48fb-b965-bf89eb95885c-kube-api-access-92mrb" (OuterVolumeSpecName: "kube-api-access-92mrb") pod "db1d9746-4bfb-48fb-b965-bf89eb95885c" (UID: "db1d9746-4bfb-48fb-b965-bf89eb95885c"). InnerVolumeSpecName "kube-api-access-92mrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:35 crc kubenswrapper[4848]: I1204 14:14:35.906007 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92mrb\" (UniqueName: \"kubernetes.io/projected/db1d9746-4bfb-48fb-b965-bf89eb95885c-kube-api-access-92mrb\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:35 crc kubenswrapper[4848]: I1204 14:14:35.906048 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db1d9746-4bfb-48fb-b965-bf89eb95885c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:36 crc kubenswrapper[4848]: I1204 14:14:36.405756 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70bfcde6-2ad2-44fc-ae39-5821c63f4918" path="/var/lib/kubelet/pods/70bfcde6-2ad2-44fc-ae39-5821c63f4918/volumes" Dec 04 14:14:36 crc kubenswrapper[4848]: I1204 14:14:36.406752 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b38b0b7c-d339-418d-b2c4-0278a451626a" path="/var/lib/kubelet/pods/b38b0b7c-d339-418d-b2c4-0278a451626a/volumes" Dec 04 14:14:36 crc kubenswrapper[4848]: I1204 14:14:36.449532 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-sl25v" event={"ID":"db1d9746-4bfb-48fb-b965-bf89eb95885c","Type":"ContainerDied","Data":"11f5f425f2d192fb11599370b0102114e94f5503507185c84005900872327141"} Dec 04 14:14:36 crc kubenswrapper[4848]: I1204 14:14:36.449583 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11f5f425f2d192fb11599370b0102114e94f5503507185c84005900872327141" Dec 04 14:14:36 crc kubenswrapper[4848]: I1204 14:14:36.449654 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-sl25v" Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.236514 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.342404 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6b33655-7b9c-4663-b8d0-85be8361f65f-operator-scripts\") pod \"d6b33655-7b9c-4663-b8d0-85be8361f65f\" (UID: \"d6b33655-7b9c-4663-b8d0-85be8361f65f\") " Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.342692 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhts5\" (UniqueName: \"kubernetes.io/projected/d6b33655-7b9c-4663-b8d0-85be8361f65f-kube-api-access-zhts5\") pod \"d6b33655-7b9c-4663-b8d0-85be8361f65f\" (UID: \"d6b33655-7b9c-4663-b8d0-85be8361f65f\") " Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.344547 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6b33655-7b9c-4663-b8d0-85be8361f65f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6b33655-7b9c-4663-b8d0-85be8361f65f" (UID: "d6b33655-7b9c-4663-b8d0-85be8361f65f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.352127 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b33655-7b9c-4663-b8d0-85be8361f65f-kube-api-access-zhts5" (OuterVolumeSpecName: "kube-api-access-zhts5") pod "d6b33655-7b9c-4663-b8d0-85be8361f65f" (UID: "d6b33655-7b9c-4663-b8d0-85be8361f65f"). InnerVolumeSpecName "kube-api-access-zhts5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.444815 4848 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6b33655-7b9c-4663-b8d0-85be8361f65f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.444838 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhts5\" (UniqueName: \"kubernetes.io/projected/d6b33655-7b9c-4663-b8d0-85be8361f65f-kube-api-access-zhts5\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.463981 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-fb6e-account-create-update-sdrw9" event={"ID":"d6b33655-7b9c-4663-b8d0-85be8361f65f","Type":"ContainerDied","Data":"759b3b4aaa784769502708de23fad28653803e95715dcae96dd9dc3ef86357d1"} Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.464024 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="759b3b4aaa784769502708de23fad28653803e95715dcae96dd9dc3ef86357d1" Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.464134 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-fb6e-account-create-update-sdrw9" Dec 04 14:14:37 crc kubenswrapper[4848]: I1204 14:14:37.730902 4848 scope.go:117] "RemoveContainer" containerID="27705653e8c1999c0d75d2368cf55c82f8fa6f3a8310aaff6145d1e1f5f73caa" Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.503765 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf","Type":"ContainerStarted","Data":"719e457daf993d730ead0c91d80dfb6abb4bf2000a8c55fab4ab255c41449ceb"} Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.514632 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" event={"ID":"576a9a23-e104-4d55-8100-ee23628af7cf","Type":"ContainerStarted","Data":"18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13"} Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.515445 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.518094 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8583482-412b-4ab5-8089-6eb3587cc0ef","Type":"ContainerStarted","Data":"a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2"} Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.518130 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8583482-412b-4ab5-8089-6eb3587cc0ef","Type":"ContainerStarted","Data":"c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa"} Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.518235 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerName="nova-metadata-log" containerID="cri-o://c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa" gracePeriod=30 Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.518309 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerName="nova-metadata-metadata" containerID="cri-o://a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2" gracePeriod=30 Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.521509 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fff65b6-de45-4e47-9224-d8620d7ae78e","Type":"ContainerStarted","Data":"2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4"} Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.521869 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5fff65b6-de45-4e47-9224-d8620d7ae78e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4" gracePeriod=30 Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.524324 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"30e84076-e3b2-4bac-b59a-9085a8f256ff","Type":"ContainerStarted","Data":"700845d450609cd769cf16f25928c4cd4ba819330e2d2d966b7ad258f57d431b"} Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.549704 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" podStartSLOduration=8.549685512 podStartE2EDuration="8.549685512s" podCreationTimestamp="2025-12-04 14:14:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:14:38.536555712 +0000 UTC m=+1582.479052260" watchObservedRunningTime="2025-12-04 14:14:38.549685512 +0000 UTC m=+1582.492182040" Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.576766 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.507127661 podStartE2EDuration="8.576746068s" podCreationTimestamp="2025-12-04 14:14:30 +0000 UTC" firstStartedPulling="2025-12-04 14:14:32.825674949 +0000 UTC m=+1576.768171477" lastFinishedPulling="2025-12-04 14:14:37.895293346 +0000 UTC m=+1581.837789884" observedRunningTime="2025-12-04 14:14:38.567336669 +0000 UTC m=+1582.509833197" watchObservedRunningTime="2025-12-04 14:14:38.576746068 +0000 UTC m=+1582.519242596" Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.610636 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.48659987 podStartE2EDuration="8.61061525s" podCreationTimestamp="2025-12-04 14:14:30 +0000 UTC" firstStartedPulling="2025-12-04 14:14:32.770772855 +0000 UTC m=+1576.713269383" lastFinishedPulling="2025-12-04 14:14:37.894788235 +0000 UTC m=+1581.837284763" observedRunningTime="2025-12-04 14:14:38.59863707 +0000 UTC m=+1582.541133598" watchObservedRunningTime="2025-12-04 14:14:38.61061525 +0000 UTC m=+1582.553111778" Dec 04 14:14:38 crc kubenswrapper[4848]: I1204 14:14:38.631147 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.554612418 podStartE2EDuration="8.631128928s" podCreationTimestamp="2025-12-04 14:14:30 +0000 UTC" firstStartedPulling="2025-12-04 14:14:31.818354527 +0000 UTC m=+1575.760851055" lastFinishedPulling="2025-12-04 14:14:37.894871037 +0000 UTC m=+1581.837367565" observedRunningTime="2025-12-04 14:14:38.615619962 +0000 UTC m=+1582.558116520" watchObservedRunningTime="2025-12-04 14:14:38.631128928 +0000 UTC m=+1582.573625456" Dec 04 14:14:39 crc kubenswrapper[4848]: I1204 14:14:39.541240 4848 generic.go:334] "Generic (PLEG): container finished" podID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerID="c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa" exitCode=143 Dec 04 14:14:39 crc kubenswrapper[4848]: I1204 14:14:39.541301 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8583482-412b-4ab5-8089-6eb3587cc0ef","Type":"ContainerDied","Data":"c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa"} Dec 04 14:14:39 crc kubenswrapper[4848]: I1204 14:14:39.544655 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf","Type":"ContainerStarted","Data":"e45e7e3b6e4dbfc2c4104d223f2dedb871d175a7b0bc45a189ac52f9ef5cfb8b"} Dec 04 14:14:39 crc kubenswrapper[4848]: I1204 14:14:39.562207 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.322205151 podStartE2EDuration="9.562167408s" podCreationTimestamp="2025-12-04 14:14:30 +0000 UTC" firstStartedPulling="2025-12-04 14:14:31.663092198 +0000 UTC m=+1575.605588726" lastFinishedPulling="2025-12-04 14:14:37.903054455 +0000 UTC m=+1581.845550983" observedRunningTime="2025-12-04 14:14:39.561218255 +0000 UTC m=+1583.503714783" watchObservedRunningTime="2025-12-04 14:14:39.562167408 +0000 UTC m=+1583.504663926" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.724458 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.724816 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.934075 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-kxmlp"] Dec 04 14:14:40 crc kubenswrapper[4848]: E1204 14:14:40.938796 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38b0b7c-d339-418d-b2c4-0278a451626a" containerName="heat-cfnapi" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.938841 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38b0b7c-d339-418d-b2c4-0278a451626a" containerName="heat-cfnapi" Dec 04 14:14:40 crc kubenswrapper[4848]: E1204 14:14:40.938870 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db1d9746-4bfb-48fb-b965-bf89eb95885c" containerName="mariadb-database-create" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.938879 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="db1d9746-4bfb-48fb-b965-bf89eb95885c" containerName="mariadb-database-create" Dec 04 14:14:40 crc kubenswrapper[4848]: E1204 14:14:40.938913 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bfcde6-2ad2-44fc-ae39-5821c63f4918" containerName="heat-api" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.938921 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bfcde6-2ad2-44fc-ae39-5821c63f4918" containerName="heat-api" Dec 04 14:14:40 crc kubenswrapper[4848]: E1204 14:14:40.938942 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b33655-7b9c-4663-b8d0-85be8361f65f" containerName="mariadb-account-create-update" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.938969 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b33655-7b9c-4663-b8d0-85be8361f65f" containerName="mariadb-account-create-update" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.939296 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="70bfcde6-2ad2-44fc-ae39-5821c63f4918" containerName="heat-api" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.939332 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b33655-7b9c-4663-b8d0-85be8361f65f" containerName="mariadb-account-create-update" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.939345 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38b0b7c-d339-418d-b2c4-0278a451626a" containerName="heat-cfnapi" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.939374 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="db1d9746-4bfb-48fb-b965-bf89eb95885c" containerName="mariadb-database-create" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.940525 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.948125 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.948294 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.948312 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-d7qhp" Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.948303 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-kxmlp"] Dec 04 14:14:40 crc kubenswrapper[4848]: I1204 14:14:40.948426 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.029516 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.029546 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.067983 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.090291 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.090368 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.131663 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-combined-ca-bundle\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.131772 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-scripts\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.132384 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-config-data\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.132522 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mpsz\" (UniqueName: \"kubernetes.io/projected/397d21bc-b1cc-49a6-b9c8-353c772cf769-kube-api-access-9mpsz\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.135510 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.234209 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mpsz\" (UniqueName: \"kubernetes.io/projected/397d21bc-b1cc-49a6-b9c8-353c772cf769-kube-api-access-9mpsz\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.234316 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-combined-ca-bundle\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.234366 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-scripts\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.234483 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-config-data\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.241561 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-combined-ca-bundle\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.242250 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-config-data\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.244611 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-scripts\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.254107 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mpsz\" (UniqueName: \"kubernetes.io/projected/397d21bc-b1cc-49a6-b9c8-353c772cf769-kube-api-access-9mpsz\") pod \"aodh-db-sync-kxmlp\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.263453 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.808171 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.233:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.808677 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.233:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.830171 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 14:14:41 crc kubenswrapper[4848]: I1204 14:14:41.839121 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-kxmlp"] Dec 04 14:14:42 crc kubenswrapper[4848]: I1204 14:14:42.796403 4848 generic.go:334] "Generic (PLEG): container finished" podID="618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" containerID="e6bb0c6f78150ddbeb78da3883863401a04086ec3f231a8cdcfeaf559006f080" exitCode=0 Dec 04 14:14:42 crc kubenswrapper[4848]: I1204 14:14:42.796583 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-s9gd9" event={"ID":"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9","Type":"ContainerDied","Data":"e6bb0c6f78150ddbeb78da3883863401a04086ec3f231a8cdcfeaf559006f080"} Dec 04 14:14:42 crc kubenswrapper[4848]: I1204 14:14:42.806370 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-kxmlp" event={"ID":"397d21bc-b1cc-49a6-b9c8-353c772cf769","Type":"ContainerStarted","Data":"51d3fd173849cfb6673b90637f4e3f459c7547ba5caa99c95be64d0da04e00ac"} Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.314506 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.314881 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.348634 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.526281 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-combined-ca-bundle\") pod \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.526801 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-config-data\") pod \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.526856 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-scripts\") pod \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.526934 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zznk2\" (UniqueName: \"kubernetes.io/projected/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-kube-api-access-zznk2\") pod \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\" (UID: \"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9\") " Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.545529 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-scripts" (OuterVolumeSpecName: "scripts") pod "618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" (UID: "618e60f3-6ba7-435e-b112-8ac2e7c6e5d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.545861 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-kube-api-access-zznk2" (OuterVolumeSpecName: "kube-api-access-zznk2") pod "618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" (UID: "618e60f3-6ba7-435e-b112-8ac2e7c6e5d9"). InnerVolumeSpecName "kube-api-access-zznk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.563615 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-config-data" (OuterVolumeSpecName: "config-data") pod "618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" (UID: "618e60f3-6ba7-435e-b112-8ac2e7c6e5d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.577081 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" (UID: "618e60f3-6ba7-435e-b112-8ac2e7c6e5d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.633510 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.633553 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.633570 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zznk2\" (UniqueName: \"kubernetes.io/projected/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-kube-api-access-zznk2\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.633584 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.863224 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-s9gd9" event={"ID":"618e60f3-6ba7-435e-b112-8ac2e7c6e5d9","Type":"ContainerDied","Data":"ba9a09732f25d902e3edd21a5826b136f5bd2e935008728575ec93de93984765"} Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.863271 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba9a09732f25d902e3edd21a5826b136f5bd2e935008728575ec93de93984765" Dec 04 14:14:44 crc kubenswrapper[4848]: I1204 14:14:44.863312 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-s9gd9" Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.047897 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.048400 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-log" containerID="cri-o://719e457daf993d730ead0c91d80dfb6abb4bf2000a8c55fab4ab255c41449ceb" gracePeriod=30 Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.048726 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-api" containerID="cri-o://e45e7e3b6e4dbfc2c4104d223f2dedb871d175a7b0bc45a189ac52f9ef5cfb8b" gracePeriod=30 Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.088999 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.089214 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="30e84076-e3b2-4bac-b59a-9085a8f256ff" containerName="nova-scheduler-scheduler" containerID="cri-o://700845d450609cd769cf16f25928c4cd4ba819330e2d2d966b7ad258f57d431b" gracePeriod=30 Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.876155 4848 generic.go:334] "Generic (PLEG): container finished" podID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerID="719e457daf993d730ead0c91d80dfb6abb4bf2000a8c55fab4ab255c41449ceb" exitCode=143 Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.876424 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf","Type":"ContainerDied","Data":"719e457daf993d730ead0c91d80dfb6abb4bf2000a8c55fab4ab255c41449ceb"} Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.880895 4848 generic.go:334] "Generic (PLEG): container finished" podID="cb391cd3-6f78-4e51-ac06-14c9420d0f06" containerID="efc73cb5821d728230b71deab4febbbdd8471088fbe6454ec39fa4a2f68e23ac" exitCode=0 Dec 04 14:14:45 crc kubenswrapper[4848]: I1204 14:14:45.880926 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" event={"ID":"cb391cd3-6f78-4e51-ac06-14c9420d0f06","Type":"ContainerDied","Data":"efc73cb5821d728230b71deab4febbbdd8471088fbe6454ec39fa4a2f68e23ac"} Dec 04 14:14:46 crc kubenswrapper[4848]: E1204 14:14:46.030784 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="700845d450609cd769cf16f25928c4cd4ba819330e2d2d966b7ad258f57d431b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:14:46 crc kubenswrapper[4848]: E1204 14:14:46.032514 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="700845d450609cd769cf16f25928c4cd4ba819330e2d2d966b7ad258f57d431b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:14:46 crc kubenswrapper[4848]: E1204 14:14:46.033971 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="700845d450609cd769cf16f25928c4cd4ba819330e2d2d966b7ad258f57d431b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:14:46 crc kubenswrapper[4848]: E1204 14:14:46.034028 4848 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="30e84076-e3b2-4bac-b59a-9085a8f256ff" containerName="nova-scheduler-scheduler" Dec 04 14:14:46 crc kubenswrapper[4848]: I1204 14:14:46.343997 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:14:46 crc kubenswrapper[4848]: I1204 14:14:46.461130 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-fhzgf"] Dec 04 14:14:46 crc kubenswrapper[4848]: I1204 14:14:46.461425 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" podUID="d58a4892-6cad-4459-b103-77bdbba3e42a" containerName="dnsmasq-dns" containerID="cri-o://32e3e0bffe950f8c8455df58282141d62ad507d863dce698cd847fca738818f0" gracePeriod=10 Dec 04 14:14:46 crc kubenswrapper[4848]: I1204 14:14:46.907863 4848 generic.go:334] "Generic (PLEG): container finished" podID="d58a4892-6cad-4459-b103-77bdbba3e42a" containerID="32e3e0bffe950f8c8455df58282141d62ad507d863dce698cd847fca738818f0" exitCode=0 Dec 04 14:14:46 crc kubenswrapper[4848]: I1204 14:14:46.908543 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" event={"ID":"d58a4892-6cad-4459-b103-77bdbba3e42a","Type":"ContainerDied","Data":"32e3e0bffe950f8c8455df58282141d62ad507d863dce698cd847fca738818f0"} Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.807649 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.926091 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-scripts\") pod \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.927231 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-combined-ca-bundle\") pod \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.927409 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-config-data\") pod \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.927585 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkzhh\" (UniqueName: \"kubernetes.io/projected/cb391cd3-6f78-4e51-ac06-14c9420d0f06-kube-api-access-dkzhh\") pod \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\" (UID: \"cb391cd3-6f78-4e51-ac06-14c9420d0f06\") " Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.940421 4848 generic.go:334] "Generic (PLEG): container finished" podID="30e84076-e3b2-4bac-b59a-9085a8f256ff" containerID="700845d450609cd769cf16f25928c4cd4ba819330e2d2d966b7ad258f57d431b" exitCode=0 Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.940498 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"30e84076-e3b2-4bac-b59a-9085a8f256ff","Type":"ContainerDied","Data":"700845d450609cd769cf16f25928c4cd4ba819330e2d2d966b7ad258f57d431b"} Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.942100 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-scripts" (OuterVolumeSpecName: "scripts") pod "cb391cd3-6f78-4e51-ac06-14c9420d0f06" (UID: "cb391cd3-6f78-4e51-ac06-14c9420d0f06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.942488 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb391cd3-6f78-4e51-ac06-14c9420d0f06-kube-api-access-dkzhh" (OuterVolumeSpecName: "kube-api-access-dkzhh") pod "cb391cd3-6f78-4e51-ac06-14c9420d0f06" (UID: "cb391cd3-6f78-4e51-ac06-14c9420d0f06"). InnerVolumeSpecName "kube-api-access-dkzhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.945801 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" event={"ID":"cb391cd3-6f78-4e51-ac06-14c9420d0f06","Type":"ContainerDied","Data":"5c9ac4b60c8d9897d70e966265a3a59e1e0f2d0222e65a90b351ccbbb6761763"} Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.945859 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c9ac4b60c8d9897d70e966265a3a59e1e0f2d0222e65a90b351ccbbb6761763" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.945934 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8tc6k" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.979594 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-config-data" (OuterVolumeSpecName: "config-data") pod "cb391cd3-6f78-4e51-ac06-14c9420d0f06" (UID: "cb391cd3-6f78-4e51-ac06-14c9420d0f06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.991528 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:14:47 crc kubenswrapper[4848]: E1204 14:14:47.992049 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb391cd3-6f78-4e51-ac06-14c9420d0f06" containerName="nova-cell1-conductor-db-sync" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.992065 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb391cd3-6f78-4e51-ac06-14c9420d0f06" containerName="nova-cell1-conductor-db-sync" Dec 04 14:14:47 crc kubenswrapper[4848]: E1204 14:14:47.992119 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" containerName="nova-manage" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.992126 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" containerName="nova-manage" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.992338 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" containerName="nova-manage" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.992359 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb391cd3-6f78-4e51-ac06-14c9420d0f06" containerName="nova-cell1-conductor-db-sync" Dec 04 14:14:47 crc kubenswrapper[4848]: I1204 14:14:47.993224 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.005163 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.008565 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb391cd3-6f78-4e51-ac06-14c9420d0f06" (UID: "cb391cd3-6f78-4e51-ac06-14c9420d0f06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.034118 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.034147 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.034162 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb391cd3-6f78-4e51-ac06-14c9420d0f06-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.034171 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkzhh\" (UniqueName: \"kubernetes.io/projected/cb391cd3-6f78-4e51-ac06-14c9420d0f06-kube-api-access-dkzhh\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.150493 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh86r\" (UniqueName: \"kubernetes.io/projected/fd31580e-d25b-49af-a7c6-2eb44a77af46-kube-api-access-bh86r\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.150901 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd31580e-d25b-49af-a7c6-2eb44a77af46-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.151051 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd31580e-d25b-49af-a7c6-2eb44a77af46-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.255446 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh86r\" (UniqueName: \"kubernetes.io/projected/fd31580e-d25b-49af-a7c6-2eb44a77af46-kube-api-access-bh86r\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.255541 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd31580e-d25b-49af-a7c6-2eb44a77af46-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.255640 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd31580e-d25b-49af-a7c6-2eb44a77af46-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.276731 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd31580e-d25b-49af-a7c6-2eb44a77af46-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.282703 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd31580e-d25b-49af-a7c6-2eb44a77af46-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.288660 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh86r\" (UniqueName: \"kubernetes.io/projected/fd31580e-d25b-49af-a7c6-2eb44a77af46-kube-api-access-bh86r\") pod \"nova-cell1-conductor-0\" (UID: \"fd31580e-d25b-49af-a7c6-2eb44a77af46\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.293818 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.325828 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.361607 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.464907 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-config\") pod \"d58a4892-6cad-4459-b103-77bdbba3e42a\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.465022 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-config-data\") pod \"30e84076-e3b2-4bac-b59a-9085a8f256ff\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.465052 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-nb\") pod \"d58a4892-6cad-4459-b103-77bdbba3e42a\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.465124 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjh89\" (UniqueName: \"kubernetes.io/projected/30e84076-e3b2-4bac-b59a-9085a8f256ff-kube-api-access-sjh89\") pod \"30e84076-e3b2-4bac-b59a-9085a8f256ff\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.465188 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8twp\" (UniqueName: \"kubernetes.io/projected/d58a4892-6cad-4459-b103-77bdbba3e42a-kube-api-access-j8twp\") pod \"d58a4892-6cad-4459-b103-77bdbba3e42a\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.465218 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-swift-storage-0\") pod \"d58a4892-6cad-4459-b103-77bdbba3e42a\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.465280 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-sb\") pod \"d58a4892-6cad-4459-b103-77bdbba3e42a\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.465388 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-svc\") pod \"d58a4892-6cad-4459-b103-77bdbba3e42a\" (UID: \"d58a4892-6cad-4459-b103-77bdbba3e42a\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.465446 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-combined-ca-bundle\") pod \"30e84076-e3b2-4bac-b59a-9085a8f256ff\" (UID: \"30e84076-e3b2-4bac-b59a-9085a8f256ff\") " Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.491429 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d58a4892-6cad-4459-b103-77bdbba3e42a-kube-api-access-j8twp" (OuterVolumeSpecName: "kube-api-access-j8twp") pod "d58a4892-6cad-4459-b103-77bdbba3e42a" (UID: "d58a4892-6cad-4459-b103-77bdbba3e42a"). InnerVolumeSpecName "kube-api-access-j8twp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:48 crc kubenswrapper[4848]: I1204 14:14:48.493973 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30e84076-e3b2-4bac-b59a-9085a8f256ff-kube-api-access-sjh89" (OuterVolumeSpecName: "kube-api-access-sjh89") pod "30e84076-e3b2-4bac-b59a-9085a8f256ff" (UID: "30e84076-e3b2-4bac-b59a-9085a8f256ff"). InnerVolumeSpecName "kube-api-access-sjh89". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.562102 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30e84076-e3b2-4bac-b59a-9085a8f256ff" (UID: "30e84076-e3b2-4bac-b59a-9085a8f256ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.570665 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-config-data" (OuterVolumeSpecName: "config-data") pod "30e84076-e3b2-4bac-b59a-9085a8f256ff" (UID: "30e84076-e3b2-4bac-b59a-9085a8f256ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.570974 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8twp\" (UniqueName: \"kubernetes.io/projected/d58a4892-6cad-4459-b103-77bdbba3e42a-kube-api-access-j8twp\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.570988 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.570997 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30e84076-e3b2-4bac-b59a-9085a8f256ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.571027 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjh89\" (UniqueName: \"kubernetes.io/projected/30e84076-e3b2-4bac-b59a-9085a8f256ff-kube-api-access-sjh89\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.591677 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d58a4892-6cad-4459-b103-77bdbba3e42a" (UID: "d58a4892-6cad-4459-b103-77bdbba3e42a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.602593 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d58a4892-6cad-4459-b103-77bdbba3e42a" (UID: "d58a4892-6cad-4459-b103-77bdbba3e42a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.602794 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d58a4892-6cad-4459-b103-77bdbba3e42a" (UID: "d58a4892-6cad-4459-b103-77bdbba3e42a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.616731 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d58a4892-6cad-4459-b103-77bdbba3e42a" (UID: "d58a4892-6cad-4459-b103-77bdbba3e42a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.624029 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-config" (OuterVolumeSpecName: "config") pod "d58a4892-6cad-4459-b103-77bdbba3e42a" (UID: "d58a4892-6cad-4459-b103-77bdbba3e42a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.673381 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.673412 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.673447 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.673461 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.673471 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d58a4892-6cad-4459-b103-77bdbba3e42a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.958549 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" event={"ID":"d58a4892-6cad-4459-b103-77bdbba3e42a","Type":"ContainerDied","Data":"d1498f5a4a349cf0a8105841a23fb6ac74fe277e758fe841de52bbd0017cfa6f"} Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.958589 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-fhzgf" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.958598 4848 scope.go:117] "RemoveContainer" containerID="32e3e0bffe950f8c8455df58282141d62ad507d863dce698cd847fca738818f0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.960455 4848 generic.go:334] "Generic (PLEG): container finished" podID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerID="e45e7e3b6e4dbfc2c4104d223f2dedb871d175a7b0bc45a189ac52f9ef5cfb8b" exitCode=0 Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.960496 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf","Type":"ContainerDied","Data":"e45e7e3b6e4dbfc2c4104d223f2dedb871d175a7b0bc45a189ac52f9ef5cfb8b"} Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.962250 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"30e84076-e3b2-4bac-b59a-9085a8f256ff","Type":"ContainerDied","Data":"6080cd82eedbcc3ade843ffa1c01d365b8ee74c279f7c3a66b00d0b99a8e40df"} Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.962329 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.967062 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-kxmlp" event={"ID":"397d21bc-b1cc-49a6-b9c8-353c772cf769","Type":"ContainerStarted","Data":"544f5b38aa92e3217e058b7d9623df4cec055b4aa04ae127917b6d64e77ac8ae"} Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:48.993312 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-kxmlp" podStartSLOduration=3.041146255 podStartE2EDuration="8.993295027s" podCreationTimestamp="2025-12-04 14:14:40 +0000 UTC" firstStartedPulling="2025-12-04 14:14:41.843717839 +0000 UTC m=+1585.786214367" lastFinishedPulling="2025-12-04 14:14:47.795866611 +0000 UTC m=+1591.738363139" observedRunningTime="2025-12-04 14:14:48.982499645 +0000 UTC m=+1592.924996173" watchObservedRunningTime="2025-12-04 14:14:48.993295027 +0000 UTC m=+1592.935791555" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.008540 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.023415 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.045150 4848 scope.go:117] "RemoveContainer" containerID="e29d436c8560a47c9e20661f65c6536970d5a6f67a0d5b31ec521f1c83361995" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.094448 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-fhzgf"] Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.105495 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-fhzgf"] Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.107134 4848 scope.go:117] "RemoveContainer" containerID="700845d450609cd769cf16f25928c4cd4ba819330e2d2d966b7ad258f57d431b" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.117023 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:49 crc kubenswrapper[4848]: E1204 14:14:49.117582 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d58a4892-6cad-4459-b103-77bdbba3e42a" containerName="dnsmasq-dns" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.117594 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d58a4892-6cad-4459-b103-77bdbba3e42a" containerName="dnsmasq-dns" Dec 04 14:14:49 crc kubenswrapper[4848]: E1204 14:14:49.117640 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e84076-e3b2-4bac-b59a-9085a8f256ff" containerName="nova-scheduler-scheduler" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.117646 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e84076-e3b2-4bac-b59a-9085a8f256ff" containerName="nova-scheduler-scheduler" Dec 04 14:14:49 crc kubenswrapper[4848]: E1204 14:14:49.117667 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d58a4892-6cad-4459-b103-77bdbba3e42a" containerName="init" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.117673 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d58a4892-6cad-4459-b103-77bdbba3e42a" containerName="init" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.117876 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="30e84076-e3b2-4bac-b59a-9085a8f256ff" containerName="nova-scheduler-scheduler" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.117892 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d58a4892-6cad-4459-b103-77bdbba3e42a" containerName="dnsmasq-dns" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.118732 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.120955 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.126396 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.183556 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.183707 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-config-data\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.183769 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtjv4\" (UniqueName: \"kubernetes.io/projected/f7c7d746-2b41-4319-9342-bdca97131d23-kube-api-access-gtjv4\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.285499 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-config-data\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.285746 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtjv4\" (UniqueName: \"kubernetes.io/projected/f7c7d746-2b41-4319-9342-bdca97131d23-kube-api-access-gtjv4\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.285879 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.300735 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-config-data\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.303798 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.314751 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtjv4\" (UniqueName: \"kubernetes.io/projected/f7c7d746-2b41-4319-9342-bdca97131d23-kube-api-access-gtjv4\") pod \"nova-scheduler-0\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " pod="openstack/nova-scheduler-0" Dec 04 14:14:49 crc kubenswrapper[4848]: I1204 14:14:49.465127 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.085708 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:14:50 crc kubenswrapper[4848]: W1204 14:14:50.097338 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd31580e_d25b_49af_a7c6_2eb44a77af46.slice/crio-04647fc42291c6f8eb26de7d6b0ce9f3b89dc1e7bb347ec71e6a5c28da1097d9 WatchSource:0}: Error finding container 04647fc42291c6f8eb26de7d6b0ce9f3b89dc1e7bb347ec71e6a5c28da1097d9: Status 404 returned error can't find the container with id 04647fc42291c6f8eb26de7d6b0ce9f3b89dc1e7bb347ec71e6a5c28da1097d9 Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.212789 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.279319 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:14:50 crc kubenswrapper[4848]: W1204 14:14:50.286266 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice/crio-00c26ea5e51202456c01c045d6e5021db0eb4121dd94459c01153924bd08e78e WatchSource:0}: Error finding container 00c26ea5e51202456c01c045d6e5021db0eb4121dd94459c01153924bd08e78e: Status 404 returned error can't find the container with id 00c26ea5e51202456c01c045d6e5021db0eb4121dd94459c01153924bd08e78e Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.308882 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-config-data\") pod \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.309138 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-logs\") pod \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.309734 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtgh9\" (UniqueName: \"kubernetes.io/projected/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-kube-api-access-rtgh9\") pod \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.309883 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-combined-ca-bundle\") pod \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\" (UID: \"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf\") " Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.311258 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-logs" (OuterVolumeSpecName: "logs") pod "84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" (UID: "84e9d1d6-5a46-4038-ad61-3c8117fa2bcf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.311502 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.314550 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-kube-api-access-rtgh9" (OuterVolumeSpecName: "kube-api-access-rtgh9") pod "84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" (UID: "84e9d1d6-5a46-4038-ad61-3c8117fa2bcf"). InnerVolumeSpecName "kube-api-access-rtgh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.349679 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-config-data" (OuterVolumeSpecName: "config-data") pod "84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" (UID: "84e9d1d6-5a46-4038-ad61-3c8117fa2bcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.352470 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" (UID: "84e9d1d6-5a46-4038-ad61-3c8117fa2bcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.410347 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30e84076-e3b2-4bac-b59a-9085a8f256ff" path="/var/lib/kubelet/pods/30e84076-e3b2-4bac-b59a-9085a8f256ff/volumes" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.411397 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d58a4892-6cad-4459-b103-77bdbba3e42a" path="/var/lib/kubelet/pods/d58a4892-6cad-4459-b103-77bdbba3e42a/volumes" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.419920 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.419981 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtgh9\" (UniqueName: \"kubernetes.io/projected/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-kube-api-access-rtgh9\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:50 crc kubenswrapper[4848]: I1204 14:14:50.419995 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.004433 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fd31580e-d25b-49af-a7c6-2eb44a77af46","Type":"ContainerStarted","Data":"cd123895884e24f750abc6b56dbc3be07bcb5e0262690ab817397de989d4d90b"} Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.004732 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.004743 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fd31580e-d25b-49af-a7c6-2eb44a77af46","Type":"ContainerStarted","Data":"04647fc42291c6f8eb26de7d6b0ce9f3b89dc1e7bb347ec71e6a5c28da1097d9"} Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.011106 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"84e9d1d6-5a46-4038-ad61-3c8117fa2bcf","Type":"ContainerDied","Data":"0bfef95fba0da49d2a8133742527d7fe09eae63b39e16394092c9693052700e5"} Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.011158 4848 scope.go:117] "RemoveContainer" containerID="e45e7e3b6e4dbfc2c4104d223f2dedb871d175a7b0bc45a189ac52f9ef5cfb8b" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.011300 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.023323 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7c7d746-2b41-4319-9342-bdca97131d23","Type":"ContainerStarted","Data":"362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665"} Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.023378 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7c7d746-2b41-4319-9342-bdca97131d23","Type":"ContainerStarted","Data":"00c26ea5e51202456c01c045d6e5021db0eb4121dd94459c01153924bd08e78e"} Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.038447 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=4.038423819 podStartE2EDuration="4.038423819s" podCreationTimestamp="2025-12-04 14:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:14:51.026658244 +0000 UTC m=+1594.969154772" watchObservedRunningTime="2025-12-04 14:14:51.038423819 +0000 UTC m=+1594.980920347" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.057568 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.057546794 podStartE2EDuration="2.057546794s" podCreationTimestamp="2025-12-04 14:14:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:14:51.045361518 +0000 UTC m=+1594.987858046" watchObservedRunningTime="2025-12-04 14:14:51.057546794 +0000 UTC m=+1595.000043322" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.068700 4848 scope.go:117] "RemoveContainer" containerID="719e457daf993d730ead0c91d80dfb6abb4bf2000a8c55fab4ab255c41449ceb" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.072908 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.088713 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.101992 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:51 crc kubenswrapper[4848]: E1204 14:14:51.102542 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-api" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.102560 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-api" Dec 04 14:14:51 crc kubenswrapper[4848]: E1204 14:14:51.102594 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-log" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.102604 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-log" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.102819 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-api" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.102842 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" containerName="nova-api-log" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.104151 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.109032 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.114430 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.235480 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.236274 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m87t6\" (UniqueName: \"kubernetes.io/projected/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-kube-api-access-m87t6\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.236503 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-logs\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.236561 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-config-data\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.339030 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-logs\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.339105 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-config-data\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.339153 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.339207 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m87t6\" (UniqueName: \"kubernetes.io/projected/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-kube-api-access-m87t6\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.339493 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-logs\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.345322 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-config-data\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.346406 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.360662 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m87t6\" (UniqueName: \"kubernetes.io/projected/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-kube-api-access-m87t6\") pod \"nova-api-0\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.428577 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:14:51 crc kubenswrapper[4848]: I1204 14:14:51.905611 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:14:52 crc kubenswrapper[4848]: I1204 14:14:52.041996 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3","Type":"ContainerStarted","Data":"0d8ed6cbcdb23009b16b99575fd8ee715bf4dd5817f9f64750dc8fc95fc97d98"} Dec 04 14:14:52 crc kubenswrapper[4848]: I1204 14:14:52.408979 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84e9d1d6-5a46-4038-ad61-3c8117fa2bcf" path="/var/lib/kubelet/pods/84e9d1d6-5a46-4038-ad61-3c8117fa2bcf/volumes" Dec 04 14:14:53 crc kubenswrapper[4848]: I1204 14:14:53.059058 4848 generic.go:334] "Generic (PLEG): container finished" podID="397d21bc-b1cc-49a6-b9c8-353c772cf769" containerID="544f5b38aa92e3217e058b7d9623df4cec055b4aa04ae127917b6d64e77ac8ae" exitCode=0 Dec 04 14:14:53 crc kubenswrapper[4848]: I1204 14:14:53.059117 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-kxmlp" event={"ID":"397d21bc-b1cc-49a6-b9c8-353c772cf769","Type":"ContainerDied","Data":"544f5b38aa92e3217e058b7d9623df4cec055b4aa04ae127917b6d64e77ac8ae"} Dec 04 14:14:53 crc kubenswrapper[4848]: I1204 14:14:53.061559 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3","Type":"ContainerStarted","Data":"dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc"} Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.106793 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3","Type":"ContainerStarted","Data":"63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694"} Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.145117 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.145044839 podStartE2EDuration="3.145044839s" podCreationTimestamp="2025-12-04 14:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:14:54.13271192 +0000 UTC m=+1598.075208518" watchObservedRunningTime="2025-12-04 14:14:54.145044839 +0000 UTC m=+1598.087541387" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.465360 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.577682 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.617602 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mpsz\" (UniqueName: \"kubernetes.io/projected/397d21bc-b1cc-49a6-b9c8-353c772cf769-kube-api-access-9mpsz\") pod \"397d21bc-b1cc-49a6-b9c8-353c772cf769\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.617727 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-scripts\") pod \"397d21bc-b1cc-49a6-b9c8-353c772cf769\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.617770 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-config-data\") pod \"397d21bc-b1cc-49a6-b9c8-353c772cf769\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.617792 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-combined-ca-bundle\") pod \"397d21bc-b1cc-49a6-b9c8-353c772cf769\" (UID: \"397d21bc-b1cc-49a6-b9c8-353c772cf769\") " Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.624079 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-scripts" (OuterVolumeSpecName: "scripts") pod "397d21bc-b1cc-49a6-b9c8-353c772cf769" (UID: "397d21bc-b1cc-49a6-b9c8-353c772cf769"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.624536 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/397d21bc-b1cc-49a6-b9c8-353c772cf769-kube-api-access-9mpsz" (OuterVolumeSpecName: "kube-api-access-9mpsz") pod "397d21bc-b1cc-49a6-b9c8-353c772cf769" (UID: "397d21bc-b1cc-49a6-b9c8-353c772cf769"). InnerVolumeSpecName "kube-api-access-9mpsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.656186 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-config-data" (OuterVolumeSpecName: "config-data") pod "397d21bc-b1cc-49a6-b9c8-353c772cf769" (UID: "397d21bc-b1cc-49a6-b9c8-353c772cf769"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.662378 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "397d21bc-b1cc-49a6-b9c8-353c772cf769" (UID: "397d21bc-b1cc-49a6-b9c8-353c772cf769"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.719853 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.719884 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.719895 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/397d21bc-b1cc-49a6-b9c8-353c772cf769-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:54 crc kubenswrapper[4848]: I1204 14:14:54.719906 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mpsz\" (UniqueName: \"kubernetes.io/projected/397d21bc-b1cc-49a6-b9c8-353c772cf769-kube-api-access-9mpsz\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:55 crc kubenswrapper[4848]: I1204 14:14:55.122080 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-kxmlp" Dec 04 14:14:55 crc kubenswrapper[4848]: I1204 14:14:55.123578 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-kxmlp" event={"ID":"397d21bc-b1cc-49a6-b9c8-353c772cf769","Type":"ContainerDied","Data":"51d3fd173849cfb6673b90637f4e3f459c7547ba5caa99c95be64d0da04e00ac"} Dec 04 14:14:55 crc kubenswrapper[4848]: I1204 14:14:55.123637 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d3fd173849cfb6673b90637f4e3f459c7547ba5caa99c95be64d0da04e00ac" Dec 04 14:14:55 crc kubenswrapper[4848]: I1204 14:14:55.229435 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.804625 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 04 14:14:57 crc kubenswrapper[4848]: E1204 14:14:57.805482 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="397d21bc-b1cc-49a6-b9c8-353c772cf769" containerName="aodh-db-sync" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.805494 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="397d21bc-b1cc-49a6-b9c8-353c772cf769" containerName="aodh-db-sync" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.805744 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="397d21bc-b1cc-49a6-b9c8-353c772cf769" containerName="aodh-db-sync" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.816811 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.819848 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.820026 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-d7qhp" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.820144 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.820541 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.890908 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntkgq\" (UniqueName: \"kubernetes.io/projected/01ba3e54-9bc1-49f4-ad14-195c52649b8c-kube-api-access-ntkgq\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.890958 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-config-data\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.891000 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-scripts\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.891047 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.993575 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntkgq\" (UniqueName: \"kubernetes.io/projected/01ba3e54-9bc1-49f4-ad14-195c52649b8c-kube-api-access-ntkgq\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.993830 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-config-data\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.993857 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-scripts\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:57 crc kubenswrapper[4848]: I1204 14:14:57.993904 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:58 crc kubenswrapper[4848]: I1204 14:14:58.001238 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:58 crc kubenswrapper[4848]: I1204 14:14:58.002795 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-config-data\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:58 crc kubenswrapper[4848]: I1204 14:14:58.007924 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-scripts\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:58 crc kubenswrapper[4848]: I1204 14:14:58.018637 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntkgq\" (UniqueName: \"kubernetes.io/projected/01ba3e54-9bc1-49f4-ad14-195c52649b8c-kube-api-access-ntkgq\") pod \"aodh-0\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " pod="openstack/aodh-0" Dec 04 14:14:58 crc kubenswrapper[4848]: I1204 14:14:58.155768 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:14:58 crc kubenswrapper[4848]: I1204 14:14:58.362176 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 14:14:58 crc kubenswrapper[4848]: W1204 14:14:58.752758 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ba3e54_9bc1_49f4_ad14_195c52649b8c.slice/crio-d3a13123ffbba2c50e0890a2581b196deb64757bca23ec5b0b177d455f35fc9e WatchSource:0}: Error finding container d3a13123ffbba2c50e0890a2581b196deb64757bca23ec5b0b177d455f35fc9e: Status 404 returned error can't find the container with id d3a13123ffbba2c50e0890a2581b196deb64757bca23ec5b0b177d455f35fc9e Dec 04 14:14:58 crc kubenswrapper[4848]: I1204 14:14:58.764207 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 14:14:59 crc kubenswrapper[4848]: I1204 14:14:59.158720 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerStarted","Data":"d3a13123ffbba2c50e0890a2581b196deb64757bca23ec5b0b177d455f35fc9e"} Dec 04 14:14:59 crc kubenswrapper[4848]: I1204 14:14:59.465651 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 14:14:59 crc kubenswrapper[4848]: I1204 14:14:59.509150 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.155142 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4"] Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.157122 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.162585 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.162605 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.170884 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4"] Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.229700 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.242453 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-config-volume\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.242519 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-secret-volume\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.242582 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtft7\" (UniqueName: \"kubernetes.io/projected/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-kube-api-access-mtft7\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.344349 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-config-volume\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.344689 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-secret-volume\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.344748 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtft7\" (UniqueName: \"kubernetes.io/projected/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-kube-api-access-mtft7\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.345783 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-config-volume\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.351846 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-secret-volume\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.369393 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtft7\" (UniqueName: \"kubernetes.io/projected/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-kube-api-access-mtft7\") pod \"collect-profiles-29414295-8kqr4\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:00 crc kubenswrapper[4848]: I1204 14:15:00.491745 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:01 crc kubenswrapper[4848]: I1204 14:15:01.429881 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:15:01 crc kubenswrapper[4848]: I1204 14:15:01.430215 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:15:01 crc kubenswrapper[4848]: I1204 14:15:01.713935 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4"] Dec 04 14:15:02 crc kubenswrapper[4848]: I1204 14:15:02.200154 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" event={"ID":"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c","Type":"ContainerStarted","Data":"3301dc7afc318e6df9b512294fc05f712510d1207e0a4b706e828c37743cbf8c"} Dec 04 14:15:02 crc kubenswrapper[4848]: I1204 14:15:02.512161 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.244:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:15:02 crc kubenswrapper[4848]: I1204 14:15:02.512194 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.244:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:15:03 crc kubenswrapper[4848]: I1204 14:15:03.223761 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:03 crc kubenswrapper[4848]: I1204 14:15:03.224262 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="ceilometer-central-agent" containerID="cri-o://77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c" gracePeriod=30 Dec 04 14:15:03 crc kubenswrapper[4848]: I1204 14:15:03.224743 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="proxy-httpd" containerID="cri-o://4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160" gracePeriod=30 Dec 04 14:15:03 crc kubenswrapper[4848]: I1204 14:15:03.224788 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="sg-core" containerID="cri-o://b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495" gracePeriod=30 Dec 04 14:15:03 crc kubenswrapper[4848]: I1204 14:15:03.224824 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="ceilometer-notification-agent" containerID="cri-o://ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e" gracePeriod=30 Dec 04 14:15:03 crc kubenswrapper[4848]: I1204 14:15:03.247260 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerStarted","Data":"a49f648dcbaf53078904342a3ca35d6f3568275890ed9e58c49a772698c70288"} Dec 04 14:15:03 crc kubenswrapper[4848]: I1204 14:15:03.248636 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" event={"ID":"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c","Type":"ContainerStarted","Data":"7c2fadc0017816f05309698f024f113ac942635367ea9f5e829ffc9fee183022"} Dec 04 14:15:03 crc kubenswrapper[4848]: I1204 14:15:03.316901 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" podStartSLOduration=3.316881455 podStartE2EDuration="3.316881455s" podCreationTimestamp="2025-12-04 14:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:03.282308155 +0000 UTC m=+1607.224804683" watchObservedRunningTime="2025-12-04 14:15:03.316881455 +0000 UTC m=+1607.259377983" Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.276513 4848 generic.go:334] "Generic (PLEG): container finished" podID="cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c" containerID="7c2fadc0017816f05309698f024f113ac942635367ea9f5e829ffc9fee183022" exitCode=0 Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.276888 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" event={"ID":"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c","Type":"ContainerDied","Data":"7c2fadc0017816f05309698f024f113ac942635367ea9f5e829ffc9fee183022"} Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.291620 4848 generic.go:334] "Generic (PLEG): container finished" podID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerID="4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160" exitCode=0 Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.291663 4848 generic.go:334] "Generic (PLEG): container finished" podID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerID="b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495" exitCode=2 Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.291675 4848 generic.go:334] "Generic (PLEG): container finished" podID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerID="77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c" exitCode=0 Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.291710 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerDied","Data":"4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160"} Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.291742 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerDied","Data":"b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495"} Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.291756 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerDied","Data":"77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c"} Dec 04 14:15:04 crc kubenswrapper[4848]: I1204 14:15:04.309883 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 04 14:15:05 crc kubenswrapper[4848]: I1204 14:15:05.852103 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.031628 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtft7\" (UniqueName: \"kubernetes.io/projected/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-kube-api-access-mtft7\") pod \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.032071 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-config-volume\") pod \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.032185 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-secret-volume\") pod \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\" (UID: \"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c\") " Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.032888 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-config-volume" (OuterVolumeSpecName: "config-volume") pod "cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c" (UID: "cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.033532 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.043065 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c" (UID: "cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.043137 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-kube-api-access-mtft7" (OuterVolumeSpecName: "kube-api-access-mtft7") pod "cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c" (UID: "cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c"). InnerVolumeSpecName "kube-api-access-mtft7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.136168 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtft7\" (UniqueName: \"kubernetes.io/projected/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-kube-api-access-mtft7\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.136210 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.318369 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerStarted","Data":"636cd099d4df528e25e0936f756ab1d111dc07ec30ca3db85abb7895f820c46d"} Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.324581 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" event={"ID":"cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c","Type":"ContainerDied","Data":"3301dc7afc318e6df9b512294fc05f712510d1207e0a4b706e828c37743cbf8c"} Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.324618 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3301dc7afc318e6df9b512294fc05f712510d1207e0a4b706e828c37743cbf8c" Dec 04 14:15:06 crc kubenswrapper[4848]: I1204 14:15:06.324669 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.322812 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.343047 4848 generic.go:334] "Generic (PLEG): container finished" podID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerID="ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e" exitCode=0 Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.343084 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerDied","Data":"ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e"} Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.343107 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.343133 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"20711d1f-8dec-41e1-920e-50603c2c7b79","Type":"ContainerDied","Data":"1144c708ff4dde3da8e4a63694c8705103a7c845a8a4b6c9e4bea32d229d1d61"} Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.343155 4848 scope.go:117] "RemoveContainer" containerID="4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.385254 4848 scope.go:117] "RemoveContainer" containerID="b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.430365 4848 scope.go:117] "RemoveContainer" containerID="ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.467707 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-sg-core-conf-yaml\") pod \"20711d1f-8dec-41e1-920e-50603c2c7b79\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.467811 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-scripts\") pod \"20711d1f-8dec-41e1-920e-50603c2c7b79\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.467933 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-log-httpd\") pod \"20711d1f-8dec-41e1-920e-50603c2c7b79\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.467981 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-run-httpd\") pod \"20711d1f-8dec-41e1-920e-50603c2c7b79\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.468508 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "20711d1f-8dec-41e1-920e-50603c2c7b79" (UID: "20711d1f-8dec-41e1-920e-50603c2c7b79"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.468454 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "20711d1f-8dec-41e1-920e-50603c2c7b79" (UID: "20711d1f-8dec-41e1-920e-50603c2c7b79"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.470033 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdhzg\" (UniqueName: \"kubernetes.io/projected/20711d1f-8dec-41e1-920e-50603c2c7b79-kube-api-access-zdhzg\") pod \"20711d1f-8dec-41e1-920e-50603c2c7b79\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.470086 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-config-data\") pod \"20711d1f-8dec-41e1-920e-50603c2c7b79\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.470136 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-combined-ca-bundle\") pod \"20711d1f-8dec-41e1-920e-50603c2c7b79\" (UID: \"20711d1f-8dec-41e1-920e-50603c2c7b79\") " Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.470817 4848 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.470835 4848 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/20711d1f-8dec-41e1-920e-50603c2c7b79-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.478922 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20711d1f-8dec-41e1-920e-50603c2c7b79-kube-api-access-zdhzg" (OuterVolumeSpecName: "kube-api-access-zdhzg") pod "20711d1f-8dec-41e1-920e-50603c2c7b79" (UID: "20711d1f-8dec-41e1-920e-50603c2c7b79"). InnerVolumeSpecName "kube-api-access-zdhzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.486117 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-scripts" (OuterVolumeSpecName: "scripts") pod "20711d1f-8dec-41e1-920e-50603c2c7b79" (UID: "20711d1f-8dec-41e1-920e-50603c2c7b79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.513355 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "20711d1f-8dec-41e1-920e-50603c2c7b79" (UID: "20711d1f-8dec-41e1-920e-50603c2c7b79"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.549425 4848 scope.go:117] "RemoveContainer" containerID="77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.576501 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdhzg\" (UniqueName: \"kubernetes.io/projected/20711d1f-8dec-41e1-920e-50603c2c7b79-kube-api-access-zdhzg\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.576524 4848 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.576555 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.636909 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20711d1f-8dec-41e1-920e-50603c2c7b79" (UID: "20711d1f-8dec-41e1-920e-50603c2c7b79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.668000 4848 scope.go:117] "RemoveContainer" containerID="4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160" Dec 04 14:15:07 crc kubenswrapper[4848]: E1204 14:15:07.668689 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160\": container with ID starting with 4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160 not found: ID does not exist" containerID="4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.668733 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160"} err="failed to get container status \"4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160\": rpc error: code = NotFound desc = could not find container \"4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160\": container with ID starting with 4bedd4d8e71d18c88833fcad014b7cb6a011ee988d6712e46db417f68f006160 not found: ID does not exist" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.668756 4848 scope.go:117] "RemoveContainer" containerID="b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495" Dec 04 14:15:07 crc kubenswrapper[4848]: E1204 14:15:07.669099 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495\": container with ID starting with b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495 not found: ID does not exist" containerID="b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.669145 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495"} err="failed to get container status \"b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495\": rpc error: code = NotFound desc = could not find container \"b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495\": container with ID starting with b785819737419ec15f60afaa7f213064b5ef325fcc9f7c6adf47602911783495 not found: ID does not exist" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.669158 4848 scope.go:117] "RemoveContainer" containerID="ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e" Dec 04 14:15:07 crc kubenswrapper[4848]: E1204 14:15:07.673439 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e\": container with ID starting with ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e not found: ID does not exist" containerID="ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.674081 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e"} err="failed to get container status \"ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e\": rpc error: code = NotFound desc = could not find container \"ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e\": container with ID starting with ec81d8165734771a993c18a0fed7f5e25c6056582295a14e6d00db5cc207a28e not found: ID does not exist" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.674489 4848 scope.go:117] "RemoveContainer" containerID="77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.674340 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-config-data" (OuterVolumeSpecName: "config-data") pod "20711d1f-8dec-41e1-920e-50603c2c7b79" (UID: "20711d1f-8dec-41e1-920e-50603c2c7b79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:07 crc kubenswrapper[4848]: E1204 14:15:07.675401 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c\": container with ID starting with 77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c not found: ID does not exist" containerID="77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.675577 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c"} err="failed to get container status \"77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c\": rpc error: code = NotFound desc = could not find container \"77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c\": container with ID starting with 77e29d475b96fb2d57f49d83ae1d786996f311f65adfbe657c04c17ad1759f8c not found: ID does not exist" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.679262 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.679293 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20711d1f-8dec-41e1-920e-50603c2c7b79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:07 crc kubenswrapper[4848]: I1204 14:15:07.980110 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.005405 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.029922 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:08 crc kubenswrapper[4848]: E1204 14:15:08.030894 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="ceilometer-central-agent" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.030918 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="ceilometer-central-agent" Dec 04 14:15:08 crc kubenswrapper[4848]: E1204 14:15:08.030962 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="ceilometer-notification-agent" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.030971 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="ceilometer-notification-agent" Dec 04 14:15:08 crc kubenswrapper[4848]: E1204 14:15:08.030996 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="proxy-httpd" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.031002 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="proxy-httpd" Dec 04 14:15:08 crc kubenswrapper[4848]: E1204 14:15:08.031025 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c" containerName="collect-profiles" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.031030 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c" containerName="collect-profiles" Dec 04 14:15:08 crc kubenswrapper[4848]: E1204 14:15:08.031038 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="sg-core" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.031044 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="sg-core" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.031248 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="sg-core" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.031266 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c" containerName="collect-profiles" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.031274 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="ceilometer-notification-agent" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.031285 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="ceilometer-central-agent" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.031320 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" containerName="proxy-httpd" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.033194 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.043356 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.043655 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.052909 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.096712 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-config-data\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.096836 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.096893 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-run-httpd\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.097048 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-scripts\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.097217 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9g8p\" (UniqueName: \"kubernetes.io/projected/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-kube-api-access-s9g8p\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.097255 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-log-httpd\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.100327 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.202149 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9g8p\" (UniqueName: \"kubernetes.io/projected/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-kube-api-access-s9g8p\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.202586 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-log-httpd\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.202666 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.202721 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-config-data\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.202757 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.202808 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-run-httpd\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.202880 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-scripts\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.205668 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-run-httpd\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.207718 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-log-httpd\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.213917 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-config-data\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.214257 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-scripts\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.216886 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.219091 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.241005 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9g8p\" (UniqueName: \"kubernetes.io/projected/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-kube-api-access-s9g8p\") pod \"ceilometer-0\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.361539 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerStarted","Data":"c70373684745f5ec0a9f579de707b989e320e7b0328e827e16b27423d32f16f8"} Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.381475 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:08 crc kubenswrapper[4848]: I1204 14:15:08.413559 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20711d1f-8dec-41e1-920e-50603c2c7b79" path="/var/lib/kubelet/pods/20711d1f-8dec-41e1-920e-50603c2c7b79/volumes" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.095015 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.117055 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.120822 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.243274 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zltlz\" (UniqueName: \"kubernetes.io/projected/c8583482-412b-4ab5-8089-6eb3587cc0ef-kube-api-access-zltlz\") pod \"c8583482-412b-4ab5-8089-6eb3587cc0ef\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.243449 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-combined-ca-bundle\") pod \"c8583482-412b-4ab5-8089-6eb3587cc0ef\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.243574 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8583482-412b-4ab5-8089-6eb3587cc0ef-logs\") pod \"c8583482-412b-4ab5-8089-6eb3587cc0ef\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.243632 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-config-data\") pod \"5fff65b6-de45-4e47-9224-d8620d7ae78e\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.243658 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-combined-ca-bundle\") pod \"5fff65b6-de45-4e47-9224-d8620d7ae78e\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.243696 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-config-data\") pod \"c8583482-412b-4ab5-8089-6eb3587cc0ef\" (UID: \"c8583482-412b-4ab5-8089-6eb3587cc0ef\") " Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.243732 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g5hb\" (UniqueName: \"kubernetes.io/projected/5fff65b6-de45-4e47-9224-d8620d7ae78e-kube-api-access-6g5hb\") pod \"5fff65b6-de45-4e47-9224-d8620d7ae78e\" (UID: \"5fff65b6-de45-4e47-9224-d8620d7ae78e\") " Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.245518 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8583482-412b-4ab5-8089-6eb3587cc0ef-logs" (OuterVolumeSpecName: "logs") pod "c8583482-412b-4ab5-8089-6eb3587cc0ef" (UID: "c8583482-412b-4ab5-8089-6eb3587cc0ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.248986 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8583482-412b-4ab5-8089-6eb3587cc0ef-kube-api-access-zltlz" (OuterVolumeSpecName: "kube-api-access-zltlz") pod "c8583482-412b-4ab5-8089-6eb3587cc0ef" (UID: "c8583482-412b-4ab5-8089-6eb3587cc0ef"). InnerVolumeSpecName "kube-api-access-zltlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.249659 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fff65b6-de45-4e47-9224-d8620d7ae78e-kube-api-access-6g5hb" (OuterVolumeSpecName: "kube-api-access-6g5hb") pod "5fff65b6-de45-4e47-9224-d8620d7ae78e" (UID: "5fff65b6-de45-4e47-9224-d8620d7ae78e"). InnerVolumeSpecName "kube-api-access-6g5hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.283747 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-config-data" (OuterVolumeSpecName: "config-data") pod "5fff65b6-de45-4e47-9224-d8620d7ae78e" (UID: "5fff65b6-de45-4e47-9224-d8620d7ae78e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.292365 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-config-data" (OuterVolumeSpecName: "config-data") pod "c8583482-412b-4ab5-8089-6eb3587cc0ef" (UID: "c8583482-412b-4ab5-8089-6eb3587cc0ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.293227 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fff65b6-de45-4e47-9224-d8620d7ae78e" (UID: "5fff65b6-de45-4e47-9224-d8620d7ae78e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.298718 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8583482-412b-4ab5-8089-6eb3587cc0ef" (UID: "c8583482-412b-4ab5-8089-6eb3587cc0ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.347204 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8583482-412b-4ab5-8089-6eb3587cc0ef-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.347242 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.347252 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fff65b6-de45-4e47-9224-d8620d7ae78e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.347263 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.347272 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g5hb\" (UniqueName: \"kubernetes.io/projected/5fff65b6-de45-4e47-9224-d8620d7ae78e-kube-api-access-6g5hb\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.347282 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zltlz\" (UniqueName: \"kubernetes.io/projected/c8583482-412b-4ab5-8089-6eb3587cc0ef-kube-api-access-zltlz\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.347341 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8583482-412b-4ab5-8089-6eb3587cc0ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.376510 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerStarted","Data":"1c38bcadac24308d808b1415936baee05edc298fffe0518b79e884694e444bca"} Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.379450 4848 generic.go:334] "Generic (PLEG): container finished" podID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerID="a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2" exitCode=137 Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.379505 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.379519 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8583482-412b-4ab5-8089-6eb3587cc0ef","Type":"ContainerDied","Data":"a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2"} Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.379543 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8583482-412b-4ab5-8089-6eb3587cc0ef","Type":"ContainerDied","Data":"4a2126b2c5abeef91f710cc4275d10ba611f571bb15951a61481014409ef1daf"} Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.379562 4848 scope.go:117] "RemoveContainer" containerID="a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.394157 4848 generic.go:334] "Generic (PLEG): container finished" podID="5fff65b6-de45-4e47-9224-d8620d7ae78e" containerID="2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4" exitCode=137 Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.394324 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.394344 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fff65b6-de45-4e47-9224-d8620d7ae78e","Type":"ContainerDied","Data":"2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4"} Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.394758 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5fff65b6-de45-4e47-9224-d8620d7ae78e","Type":"ContainerDied","Data":"f9ac3e779da587ed1e411554738d5b21a544bab4f7d0c041205749ec343677f4"} Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.442057 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.479859 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.505223 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: E1204 14:15:09.505756 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerName="nova-metadata-log" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.505768 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerName="nova-metadata-log" Dec 04 14:15:09 crc kubenswrapper[4848]: E1204 14:15:09.505794 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fff65b6-de45-4e47-9224-d8620d7ae78e" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.505800 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fff65b6-de45-4e47-9224-d8620d7ae78e" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 14:15:09 crc kubenswrapper[4848]: E1204 14:15:09.505830 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerName="nova-metadata-metadata" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.505836 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerName="nova-metadata-metadata" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.506086 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fff65b6-de45-4e47-9224-d8620d7ae78e" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.506097 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerName="nova-metadata-metadata" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.506114 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" containerName="nova-metadata-log" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.507633 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.510172 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.510358 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.529668 4848 scope.go:117] "RemoveContainer" containerID="c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.556016 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.575831 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.602704 4848 scope.go:117] "RemoveContainer" containerID="a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.604067 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: E1204 14:15:09.607083 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2\": container with ID starting with a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2 not found: ID does not exist" containerID="a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.607135 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2"} err="failed to get container status \"a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2\": rpc error: code = NotFound desc = could not find container \"a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2\": container with ID starting with a70714e446260d0365c42aac865099d4467de746dce86f7d08eebd2f99a67ef2 not found: ID does not exist" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.607167 4848 scope.go:117] "RemoveContainer" containerID="c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa" Dec 04 14:15:09 crc kubenswrapper[4848]: E1204 14:15:09.607551 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa\": container with ID starting with c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa not found: ID does not exist" containerID="c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.607601 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa"} err="failed to get container status \"c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa\": rpc error: code = NotFound desc = could not find container \"c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa\": container with ID starting with c03b8b1977534fa6609b02dfd1a087b2b9d9896b5ced9da1205d1654153ca7aa not found: ID does not exist" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.607898 4848 scope.go:117] "RemoveContainer" containerID="2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.616774 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.618559 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.622785 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.622806 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.623064 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.627714 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.646192 4848 scope.go:117] "RemoveContainer" containerID="2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4" Dec 04 14:15:09 crc kubenswrapper[4848]: E1204 14:15:09.646606 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4\": container with ID starting with 2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4 not found: ID does not exist" containerID="2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.646679 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4"} err="failed to get container status \"2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4\": rpc error: code = NotFound desc = could not find container \"2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4\": container with ID starting with 2fbf6388b0f776830b7eabb0a0a492f80932c73139542d7d2c135f97176d3ab4 not found: ID does not exist" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.654410 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1846d8fa-57ed-495f-aca9-87369c79279b-logs\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.654465 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.654621 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-config-data\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.654783 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmx8r\" (UniqueName: \"kubernetes.io/projected/1846d8fa-57ed-495f-aca9-87369c79279b-kube-api-access-fmx8r\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.654817 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757310 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1846d8fa-57ed-495f-aca9-87369c79279b-logs\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757357 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757418 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-config-data\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757463 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmx8r\" (UniqueName: \"kubernetes.io/projected/1846d8fa-57ed-495f-aca9-87369c79279b-kube-api-access-fmx8r\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757483 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757513 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757531 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5jvz\" (UniqueName: \"kubernetes.io/projected/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-kube-api-access-f5jvz\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757552 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757573 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.757625 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.758396 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1846d8fa-57ed-495f-aca9-87369c79279b-logs\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.763824 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-config-data\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.764614 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.765751 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.781691 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmx8r\" (UniqueName: \"kubernetes.io/projected/1846d8fa-57ed-495f-aca9-87369c79279b-kube-api-access-fmx8r\") pod \"nova-metadata-0\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.832287 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.860457 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.860510 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5jvz\" (UniqueName: \"kubernetes.io/projected/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-kube-api-access-f5jvz\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.860542 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.860571 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.860632 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.864801 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.865129 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.865217 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.865288 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.876067 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5jvz\" (UniqueName: \"kubernetes.io/projected/1c49601a-f35b-41cb-bbb6-d7a4dd0542c0-kube-api-access-f5jvz\") pod \"nova-cell1-novncproxy-0\" (UID: \"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:09 crc kubenswrapper[4848]: I1204 14:15:09.941236 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.415820 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fff65b6-de45-4e47-9224-d8620d7ae78e" path="/var/lib/kubelet/pods/5fff65b6-de45-4e47-9224-d8620d7ae78e/volumes" Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.417059 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8583482-412b-4ab5-8089-6eb3587cc0ef" path="/var/lib/kubelet/pods/c8583482-412b-4ab5-8089-6eb3587cc0ef/volumes" Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.417780 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerStarted","Data":"6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a"} Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.418704 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerStarted","Data":"cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f"} Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.418836 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-api" containerID="cri-o://a49f648dcbaf53078904342a3ca35d6f3568275890ed9e58c49a772698c70288" gracePeriod=30 Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.420599 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-listener" containerID="cri-o://cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f" gracePeriod=30 Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.420688 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-notifier" containerID="cri-o://c70373684745f5ec0a9f579de707b989e320e7b0328e827e16b27423d32f16f8" gracePeriod=30 Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.420804 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-evaluator" containerID="cri-o://636cd099d4df528e25e0936f756ab1d111dc07ec30ca3db85abb7895f820c46d" gracePeriod=30 Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.478226 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.071906347 podStartE2EDuration="13.478201317s" podCreationTimestamp="2025-12-04 14:14:57 +0000 UTC" firstStartedPulling="2025-12-04 14:14:58.756053426 +0000 UTC m=+1602.698549954" lastFinishedPulling="2025-12-04 14:15:09.162348396 +0000 UTC m=+1613.104844924" observedRunningTime="2025-12-04 14:15:10.455643719 +0000 UTC m=+1614.398140247" watchObservedRunningTime="2025-12-04 14:15:10.478201317 +0000 UTC m=+1614.420697835" Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.517766 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:10 crc kubenswrapper[4848]: I1204 14:15:10.527160 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.434809 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.436146 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.438297 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.438838 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.442087 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.476229 4848 generic.go:334] "Generic (PLEG): container finished" podID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerID="c70373684745f5ec0a9f579de707b989e320e7b0328e827e16b27423d32f16f8" exitCode=0 Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.476268 4848 generic.go:334] "Generic (PLEG): container finished" podID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerID="636cd099d4df528e25e0936f756ab1d111dc07ec30ca3db85abb7895f820c46d" exitCode=0 Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.476277 4848 generic.go:334] "Generic (PLEG): container finished" podID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerID="a49f648dcbaf53078904342a3ca35d6f3568275890ed9e58c49a772698c70288" exitCode=0 Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.476319 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerDied","Data":"c70373684745f5ec0a9f579de707b989e320e7b0328e827e16b27423d32f16f8"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.476345 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerDied","Data":"636cd099d4df528e25e0936f756ab1d111dc07ec30ca3db85abb7895f820c46d"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.476353 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerDied","Data":"a49f648dcbaf53078904342a3ca35d6f3568275890ed9e58c49a772698c70288"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.477273 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.480248 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerStarted","Data":"46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.483622 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0","Type":"ContainerStarted","Data":"a544790871d283b8dd63aab44bb56e1e3d339094f22f475be1fc96f467b77c8d"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.483661 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1c49601a-f35b-41cb-bbb6-d7a4dd0542c0","Type":"ContainerStarted","Data":"5d1fe5e1bd9728e182bb27a43f2da045c799538e41ceec1ff82cae0d97701eb1"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.486402 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1846d8fa-57ed-495f-aca9-87369c79279b","Type":"ContainerStarted","Data":"bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.486431 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1846d8fa-57ed-495f-aca9-87369c79279b","Type":"ContainerStarted","Data":"592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.486442 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1846d8fa-57ed-495f-aca9-87369c79279b","Type":"ContainerStarted","Data":"6fc76e2463c5817b3dfb2c467b555d334d26697d47334ff568899049694b7372"} Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.542238 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.5422189939999997 podStartE2EDuration="2.542218994s" podCreationTimestamp="2025-12-04 14:15:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:11.524130176 +0000 UTC m=+1615.466626714" watchObservedRunningTime="2025-12-04 14:15:11.542218994 +0000 UTC m=+1615.484715522" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.566208 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.566185997 podStartE2EDuration="2.566185997s" podCreationTimestamp="2025-12-04 14:15:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:11.545628787 +0000 UTC m=+1615.488125345" watchObservedRunningTime="2025-12-04 14:15:11.566185997 +0000 UTC m=+1615.508682525" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.669321 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x"] Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.672677 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.695337 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x"] Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.818576 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fg84\" (UniqueName: \"kubernetes.io/projected/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-kube-api-access-9fg84\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.818627 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.818708 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.818747 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.818784 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.818806 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-config\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.920838 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fg84\" (UniqueName: \"kubernetes.io/projected/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-kube-api-access-9fg84\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.921146 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.921271 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.921387 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.921507 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.921586 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-config\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.922088 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.922130 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.922694 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.922696 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.923257 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-config\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:11 crc kubenswrapper[4848]: I1204 14:15:11.951473 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fg84\" (UniqueName: \"kubernetes.io/projected/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-kube-api-access-9fg84\") pod \"dnsmasq-dns-6b7bbf7cf9-6xv4x\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:12 crc kubenswrapper[4848]: I1204 14:15:12.004338 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:12 crc kubenswrapper[4848]: I1204 14:15:12.522402 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerStarted","Data":"149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c"} Dec 04 14:15:12 crc kubenswrapper[4848]: I1204 14:15:12.605051 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x"] Dec 04 14:15:13 crc kubenswrapper[4848]: I1204 14:15:13.564290 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerStarted","Data":"d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376"} Dec 04 14:15:13 crc kubenswrapper[4848]: I1204 14:15:13.566429 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:15:13 crc kubenswrapper[4848]: I1204 14:15:13.576216 4848 generic.go:334] "Generic (PLEG): container finished" podID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" containerID="c4be3ebf0c729a807282a42f023525b83585e04a0484d49371a2bc772dd9e6c6" exitCode=0 Dec 04 14:15:13 crc kubenswrapper[4848]: I1204 14:15:13.576263 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" event={"ID":"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345","Type":"ContainerDied","Data":"c4be3ebf0c729a807282a42f023525b83585e04a0484d49371a2bc772dd9e6c6"} Dec 04 14:15:13 crc kubenswrapper[4848]: I1204 14:15:13.576287 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" event={"ID":"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345","Type":"ContainerStarted","Data":"877f95dea8816ac327f1f68bb1b39017c4b5b41c6b1dff2f9b83a495869f9eca"} Dec 04 14:15:13 crc kubenswrapper[4848]: I1204 14:15:13.603257 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.612445581 podStartE2EDuration="5.603234163s" podCreationTimestamp="2025-12-04 14:15:08 +0000 UTC" firstStartedPulling="2025-12-04 14:15:09.161338301 +0000 UTC m=+1613.103834829" lastFinishedPulling="2025-12-04 14:15:13.152126873 +0000 UTC m=+1617.094623411" observedRunningTime="2025-12-04 14:15:13.589409798 +0000 UTC m=+1617.531906346" watchObservedRunningTime="2025-12-04 14:15:13.603234163 +0000 UTC m=+1617.545730691" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.313763 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.314127 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.314177 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.314822 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.314886 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" gracePeriod=600 Dec 04 14:15:14 crc kubenswrapper[4848]: E1204 14:15:14.491366 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.590498 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" exitCode=0 Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.590589 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357"} Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.590661 4848 scope.go:117] "RemoveContainer" containerID="31d6948afbef92b12725ad2b8756f5336adda03c65024c1864ef7ff8a0de5eb8" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.591503 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:15:14 crc kubenswrapper[4848]: E1204 14:15:14.591860 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.603040 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" event={"ID":"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345","Type":"ContainerStarted","Data":"5dff0ba863b6b1e82c544163b5adc8e9eeb466523f3f4d479c9f4c39e21e0c95"} Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.603080 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.724176 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" podStartSLOduration=3.724157082 podStartE2EDuration="3.724157082s" podCreationTimestamp="2025-12-04 14:15:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:14.634543747 +0000 UTC m=+1618.577040275" watchObservedRunningTime="2025-12-04 14:15:14.724157082 +0000 UTC m=+1618.666653610" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.728469 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.832548 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.832591 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.864074 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.864410 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-api" containerID="cri-o://63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694" gracePeriod=30 Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.864588 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-log" containerID="cri-o://dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc" gracePeriod=30 Dec 04 14:15:14 crc kubenswrapper[4848]: I1204 14:15:14.942414 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:15 crc kubenswrapper[4848]: I1204 14:15:15.612752 4848 generic.go:334] "Generic (PLEG): container finished" podID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerID="dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc" exitCode=143 Dec 04 14:15:15 crc kubenswrapper[4848]: I1204 14:15:15.612829 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3","Type":"ContainerDied","Data":"dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc"} Dec 04 14:15:15 crc kubenswrapper[4848]: I1204 14:15:15.616196 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="ceilometer-central-agent" containerID="cri-o://6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a" gracePeriod=30 Dec 04 14:15:15 crc kubenswrapper[4848]: I1204 14:15:15.616850 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="proxy-httpd" containerID="cri-o://d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376" gracePeriod=30 Dec 04 14:15:15 crc kubenswrapper[4848]: I1204 14:15:15.617052 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="sg-core" containerID="cri-o://149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c" gracePeriod=30 Dec 04 14:15:15 crc kubenswrapper[4848]: I1204 14:15:15.617091 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="ceilometer-notification-agent" containerID="cri-o://46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3" gracePeriod=30 Dec 04 14:15:16 crc kubenswrapper[4848]: I1204 14:15:16.637900 4848 generic.go:334] "Generic (PLEG): container finished" podID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerID="d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376" exitCode=0 Dec 04 14:15:16 crc kubenswrapper[4848]: I1204 14:15:16.637934 4848 generic.go:334] "Generic (PLEG): container finished" podID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerID="149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c" exitCode=2 Dec 04 14:15:16 crc kubenswrapper[4848]: I1204 14:15:16.637943 4848 generic.go:334] "Generic (PLEG): container finished" podID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerID="46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3" exitCode=0 Dec 04 14:15:16 crc kubenswrapper[4848]: I1204 14:15:16.637978 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerDied","Data":"d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376"} Dec 04 14:15:16 crc kubenswrapper[4848]: I1204 14:15:16.638004 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerDied","Data":"149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c"} Dec 04 14:15:16 crc kubenswrapper[4848]: I1204 14:15:16.638015 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerDied","Data":"46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3"} Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.572004 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.633541 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-config-data\") pod \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.633662 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-logs\") pod \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.633764 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-combined-ca-bundle\") pod \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.633800 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m87t6\" (UniqueName: \"kubernetes.io/projected/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-kube-api-access-m87t6\") pod \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\" (UID: \"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3\") " Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.634217 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-logs" (OuterVolumeSpecName: "logs") pod "ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" (UID: "ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.634798 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.645462 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-kube-api-access-m87t6" (OuterVolumeSpecName: "kube-api-access-m87t6") pod "ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" (UID: "ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3"). InnerVolumeSpecName "kube-api-access-m87t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.680079 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" (UID: "ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.682646 4848 generic.go:334] "Generic (PLEG): container finished" podID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerID="63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694" exitCode=0 Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.682692 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3","Type":"ContainerDied","Data":"63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694"} Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.682725 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3","Type":"ContainerDied","Data":"0d8ed6cbcdb23009b16b99575fd8ee715bf4dd5817f9f64750dc8fc95fc97d98"} Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.682745 4848 scope.go:117] "RemoveContainer" containerID="63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.682913 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.689919 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-config-data" (OuterVolumeSpecName: "config-data") pod "ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" (UID: "ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.737989 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.738032 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m87t6\" (UniqueName: \"kubernetes.io/projected/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-kube-api-access-m87t6\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.738051 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.796181 4848 scope.go:117] "RemoveContainer" containerID="dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.817281 4848 scope.go:117] "RemoveContainer" containerID="63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694" Dec 04 14:15:18 crc kubenswrapper[4848]: E1204 14:15:18.817717 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694\": container with ID starting with 63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694 not found: ID does not exist" containerID="63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.817774 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694"} err="failed to get container status \"63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694\": rpc error: code = NotFound desc = could not find container \"63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694\": container with ID starting with 63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694 not found: ID does not exist" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.817809 4848 scope.go:117] "RemoveContainer" containerID="dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc" Dec 04 14:15:18 crc kubenswrapper[4848]: E1204 14:15:18.818166 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc\": container with ID starting with dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc not found: ID does not exist" containerID="dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc" Dec 04 14:15:18 crc kubenswrapper[4848]: I1204 14:15:18.818199 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc"} err="failed to get container status \"dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc\": rpc error: code = NotFound desc = could not find container \"dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc\": container with ID starting with dc697ca4d8722b1ca20d123561d03bb37d19db51e4347564920306270265f4bc not found: ID does not exist" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.042293 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.071386 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.086839 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:19 crc kubenswrapper[4848]: E1204 14:15:19.087328 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-api" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.087339 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-api" Dec 04 14:15:19 crc kubenswrapper[4848]: E1204 14:15:19.087369 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-log" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.087376 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-log" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.087623 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-api" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.087649 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" containerName="nova-api-log" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.088804 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.098873 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.099314 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.099575 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.109669 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.149830 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.149897 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9487\" (UniqueName: \"kubernetes.io/projected/b3fbac58-059d-4bab-b329-aa13c7a4552c-kube-api-access-n9487\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.149948 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-public-tls-certs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.150026 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-config-data\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.150160 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.150190 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3fbac58-059d-4bab-b329-aa13c7a4552c-logs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.252691 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.252749 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3fbac58-059d-4bab-b329-aa13c7a4552c-logs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.252797 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.252853 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9487\" (UniqueName: \"kubernetes.io/projected/b3fbac58-059d-4bab-b329-aa13c7a4552c-kube-api-access-n9487\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.252923 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-public-tls-certs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.253093 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-config-data\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.255839 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3fbac58-059d-4bab-b329-aa13c7a4552c-logs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.259577 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.259928 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.265682 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-public-tls-certs\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.274395 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9487\" (UniqueName: \"kubernetes.io/projected/b3fbac58-059d-4bab-b329-aa13c7a4552c-kube-api-access-n9487\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.292654 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-config-data\") pod \"nova-api-0\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.453986 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.484567 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.562825 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-config-data\") pod \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.562908 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-scripts\") pod \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.562986 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-combined-ca-bundle\") pod \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.563033 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9g8p\" (UniqueName: \"kubernetes.io/projected/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-kube-api-access-s9g8p\") pod \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.563111 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-run-httpd\") pod \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.563127 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-log-httpd\") pod \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.563243 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-sg-core-conf-yaml\") pod \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\" (UID: \"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef\") " Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.563929 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" (UID: "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.564047 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" (UID: "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.569153 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-kube-api-access-s9g8p" (OuterVolumeSpecName: "kube-api-access-s9g8p") pod "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" (UID: "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef"). InnerVolumeSpecName "kube-api-access-s9g8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.578400 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-scripts" (OuterVolumeSpecName: "scripts") pod "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" (UID: "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.621711 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" (UID: "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.666870 4848 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.666899 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.666909 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9g8p\" (UniqueName: \"kubernetes.io/projected/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-kube-api-access-s9g8p\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.666919 4848 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.666927 4848 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.683461 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" (UID: "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.701499 4848 generic.go:334] "Generic (PLEG): container finished" podID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerID="6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a" exitCode=0 Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.701538 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerDied","Data":"6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a"} Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.701563 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d0d74c53-ba76-41a2-a258-a0e45f3dc2ef","Type":"ContainerDied","Data":"1c38bcadac24308d808b1415936baee05edc298fffe0518b79e884694e444bca"} Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.701578 4848 scope.go:117] "RemoveContainer" containerID="d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.701616 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.728573 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-config-data" (OuterVolumeSpecName: "config-data") pod "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" (UID: "d0d74c53-ba76-41a2-a258-a0e45f3dc2ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.745864 4848 scope.go:117] "RemoveContainer" containerID="149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.769531 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.769715 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.778251 4848 scope.go:117] "RemoveContainer" containerID="46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.806904 4848 scope.go:117] "RemoveContainer" containerID="6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.832962 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.833258 4848 scope.go:117] "RemoveContainer" containerID="d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.833436 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:15:19 crc kubenswrapper[4848]: E1204 14:15:19.834185 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376\": container with ID starting with d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376 not found: ID does not exist" containerID="d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.834233 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376"} err="failed to get container status \"d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376\": rpc error: code = NotFound desc = could not find container \"d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376\": container with ID starting with d5d1dccf49f730d471f950132f510fc1e11b25413ea1c6ec8d6b2ecb49c49376 not found: ID does not exist" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.834261 4848 scope.go:117] "RemoveContainer" containerID="149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c" Dec 04 14:15:19 crc kubenswrapper[4848]: E1204 14:15:19.834582 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c\": container with ID starting with 149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c not found: ID does not exist" containerID="149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.834619 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c"} err="failed to get container status \"149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c\": rpc error: code = NotFound desc = could not find container \"149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c\": container with ID starting with 149b619b94fcc59389700ff866a551059c5e333b9df90a7bc5c0f6d45a53485c not found: ID does not exist" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.834641 4848 scope.go:117] "RemoveContainer" containerID="46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3" Dec 04 14:15:19 crc kubenswrapper[4848]: E1204 14:15:19.834870 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3\": container with ID starting with 46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3 not found: ID does not exist" containerID="46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.834904 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3"} err="failed to get container status \"46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3\": rpc error: code = NotFound desc = could not find container \"46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3\": container with ID starting with 46bfab57d5ee490fcd2bb9cff4f93a3ad19c75618e18ccdb46ccb5ca0bc837c3 not found: ID does not exist" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.834925 4848 scope.go:117] "RemoveContainer" containerID="6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a" Dec 04 14:15:19 crc kubenswrapper[4848]: E1204 14:15:19.835202 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a\": container with ID starting with 6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a not found: ID does not exist" containerID="6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.835239 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a"} err="failed to get container status \"6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a\": rpc error: code = NotFound desc = could not find container \"6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a\": container with ID starting with 6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a not found: ID does not exist" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.941510 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:19 crc kubenswrapper[4848]: I1204 14:15:19.959801 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.095489 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.133470 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.146793 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.167978 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:20 crc kubenswrapper[4848]: E1204 14:15:20.168547 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="sg-core" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.168564 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="sg-core" Dec 04 14:15:20 crc kubenswrapper[4848]: E1204 14:15:20.168586 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="ceilometer-notification-agent" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.168593 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="ceilometer-notification-agent" Dec 04 14:15:20 crc kubenswrapper[4848]: E1204 14:15:20.168605 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="proxy-httpd" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.168611 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="proxy-httpd" Dec 04 14:15:20 crc kubenswrapper[4848]: E1204 14:15:20.168627 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="ceilometer-central-agent" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.168633 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="ceilometer-central-agent" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.168893 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="sg-core" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.168911 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="ceilometer-central-agent" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.168934 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="proxy-httpd" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.169081 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" containerName="ceilometer-notification-agent" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.171984 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.175142 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.175629 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.206958 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.305368 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.305734 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-scripts\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.305770 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kbxf\" (UniqueName: \"kubernetes.io/projected/5c8022e0-1093-4739-84b9-c8f77b8c1dea-kube-api-access-7kbxf\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.305801 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.305856 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-config-data\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.306454 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-run-httpd\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.306491 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-log-httpd\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.408769 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d74c53-ba76-41a2-a258-a0e45f3dc2ef" path="/var/lib/kubelet/pods/d0d74c53-ba76-41a2-a258-a0e45f3dc2ef/volumes" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.409358 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.409401 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-scripts\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.409444 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kbxf\" (UniqueName: \"kubernetes.io/projected/5c8022e0-1093-4739-84b9-c8f77b8c1dea-kube-api-access-7kbxf\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.409474 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.409514 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-config-data\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.409603 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3" path="/var/lib/kubelet/pods/ee226d6d-c7ec-4125-bdf4-c96c3ec0dfb3/volumes" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.409688 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-run-httpd\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.409710 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-log-httpd\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.410281 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-run-httpd\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.410431 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-log-httpd\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.417816 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.418014 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.424411 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-config-data\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.425038 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-scripts\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.426715 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kbxf\" (UniqueName: \"kubernetes.io/projected/5c8022e0-1093-4739-84b9-c8f77b8c1dea-kube-api-access-7kbxf\") pod \"ceilometer-0\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.654317 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.724574 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3fbac58-059d-4bab-b329-aa13c7a4552c","Type":"ContainerStarted","Data":"a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a"} Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.724927 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3fbac58-059d-4bab-b329-aa13c7a4552c","Type":"ContainerStarted","Data":"4ff1adfa7e2d3615dbba8d77021be244a62dc0281dfbccac220fe4748a434338"} Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.745861 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.867215 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:15:20 crc kubenswrapper[4848]: I1204 14:15:20.867487 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.092468 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-msdsj"] Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.094488 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.098731 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.098968 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.107426 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-msdsj"] Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.244578 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-scripts\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.244674 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.244803 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-config-data\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.244823 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwmgr\" (UniqueName: \"kubernetes.io/projected/1607ed31-5235-4f0c-a8e9-90f428f1da21-kube-api-access-vwmgr\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.256090 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.346968 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-config-data\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.347020 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwmgr\" (UniqueName: \"kubernetes.io/projected/1607ed31-5235-4f0c-a8e9-90f428f1da21-kube-api-access-vwmgr\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.347089 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-scripts\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.347191 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.353786 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.353830 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-config-data\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.363491 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-scripts\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.367384 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwmgr\" (UniqueName: \"kubernetes.io/projected/1607ed31-5235-4f0c-a8e9-90f428f1da21-kube-api-access-vwmgr\") pod \"nova-cell1-cell-mapping-msdsj\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.507879 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.743445 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerStarted","Data":"6ad36aeacbb5b045bb63f344c0fdb0caaa1806c3966b01fc82ea994cf09afa33"} Dec 04 14:15:21 crc kubenswrapper[4848]: I1204 14:15:21.773650 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3fbac58-059d-4bab-b329-aa13c7a4552c","Type":"ContainerStarted","Data":"1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377"} Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.006119 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.044136 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.044110854 podStartE2EDuration="3.044110854s" podCreationTimestamp="2025-12-04 14:15:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:21.79636249 +0000 UTC m=+1625.738859018" watchObservedRunningTime="2025-12-04 14:15:22.044110854 +0000 UTC m=+1625.986607382" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.045513 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-msdsj"] Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.088722 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-rx28f"] Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.089209 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" podUID="576a9a23-e104-4d55-8100-ee23628af7cf" containerName="dnsmasq-dns" containerID="cri-o://18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13" gracePeriod=10 Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.635640 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.789550 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerStarted","Data":"2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91"} Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.789610 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerStarted","Data":"c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33"} Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.790728 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-msdsj" event={"ID":"1607ed31-5235-4f0c-a8e9-90f428f1da21","Type":"ContainerStarted","Data":"1a0303464165f9c664f35f880812155bdf885db86f9276f9bb584f140fe87c8a"} Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.790754 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-msdsj" event={"ID":"1607ed31-5235-4f0c-a8e9-90f428f1da21","Type":"ContainerStarted","Data":"84b377bc08d8914b37119cef1e3a414e8e199dd16c789f147834dfce68f59c85"} Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.801539 4848 generic.go:334] "Generic (PLEG): container finished" podID="576a9a23-e104-4d55-8100-ee23628af7cf" containerID="18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13" exitCode=0 Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.802388 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" event={"ID":"576a9a23-e104-4d55-8100-ee23628af7cf","Type":"ContainerDied","Data":"18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13"} Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.802469 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.802523 4848 scope.go:117] "RemoveContainer" containerID="18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.802487 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-rx28f" event={"ID":"576a9a23-e104-4d55-8100-ee23628af7cf","Type":"ContainerDied","Data":"547e7e3cd3211e334c27881428a32645df5ac620fe0df7250b0de191e7bfe063"} Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.827581 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-msdsj" podStartSLOduration=1.827556151 podStartE2EDuration="1.827556151s" podCreationTimestamp="2025-12-04 14:15:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:22.812844614 +0000 UTC m=+1626.755341152" watchObservedRunningTime="2025-12-04 14:15:22.827556151 +0000 UTC m=+1626.770052679" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.834525 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-config\") pod \"576a9a23-e104-4d55-8100-ee23628af7cf\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.834634 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-svc\") pod \"576a9a23-e104-4d55-8100-ee23628af7cf\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.834655 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-sb\") pod \"576a9a23-e104-4d55-8100-ee23628af7cf\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.834838 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-nb\") pod \"576a9a23-e104-4d55-8100-ee23628af7cf\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.834916 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzgm8\" (UniqueName: \"kubernetes.io/projected/576a9a23-e104-4d55-8100-ee23628af7cf-kube-api-access-pzgm8\") pod \"576a9a23-e104-4d55-8100-ee23628af7cf\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.834968 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-swift-storage-0\") pod \"576a9a23-e104-4d55-8100-ee23628af7cf\" (UID: \"576a9a23-e104-4d55-8100-ee23628af7cf\") " Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.839360 4848 scope.go:117] "RemoveContainer" containerID="e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.840103 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/576a9a23-e104-4d55-8100-ee23628af7cf-kube-api-access-pzgm8" (OuterVolumeSpecName: "kube-api-access-pzgm8") pod "576a9a23-e104-4d55-8100-ee23628af7cf" (UID: "576a9a23-e104-4d55-8100-ee23628af7cf"). InnerVolumeSpecName "kube-api-access-pzgm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.864106 4848 scope.go:117] "RemoveContainer" containerID="18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13" Dec 04 14:15:22 crc kubenswrapper[4848]: E1204 14:15:22.867226 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13\": container with ID starting with 18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13 not found: ID does not exist" containerID="18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.867280 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13"} err="failed to get container status \"18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13\": rpc error: code = NotFound desc = could not find container \"18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13\": container with ID starting with 18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13 not found: ID does not exist" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.867315 4848 scope.go:117] "RemoveContainer" containerID="e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7" Dec 04 14:15:22 crc kubenswrapper[4848]: E1204 14:15:22.867878 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7\": container with ID starting with e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7 not found: ID does not exist" containerID="e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.867996 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7"} err="failed to get container status \"e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7\": rpc error: code = NotFound desc = could not find container \"e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7\": container with ID starting with e6d58749ce08db09a13517cb95276e4add9a4c1694bd7f9884642450440b85c7 not found: ID does not exist" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.897669 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "576a9a23-e104-4d55-8100-ee23628af7cf" (UID: "576a9a23-e104-4d55-8100-ee23628af7cf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.913199 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "576a9a23-e104-4d55-8100-ee23628af7cf" (UID: "576a9a23-e104-4d55-8100-ee23628af7cf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.917906 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "576a9a23-e104-4d55-8100-ee23628af7cf" (UID: "576a9a23-e104-4d55-8100-ee23628af7cf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.919598 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "576a9a23-e104-4d55-8100-ee23628af7cf" (UID: "576a9a23-e104-4d55-8100-ee23628af7cf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.940223 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzgm8\" (UniqueName: \"kubernetes.io/projected/576a9a23-e104-4d55-8100-ee23628af7cf-kube-api-access-pzgm8\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.940927 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.940961 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.940971 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.940981 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:22 crc kubenswrapper[4848]: I1204 14:15:22.948025 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-config" (OuterVolumeSpecName: "config") pod "576a9a23-e104-4d55-8100-ee23628af7cf" (UID: "576a9a23-e104-4d55-8100-ee23628af7cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:15:23 crc kubenswrapper[4848]: I1204 14:15:23.043263 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/576a9a23-e104-4d55-8100-ee23628af7cf-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:23 crc kubenswrapper[4848]: I1204 14:15:23.381220 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-rx28f"] Dec 04 14:15:23 crc kubenswrapper[4848]: I1204 14:15:23.392540 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-rx28f"] Dec 04 14:15:23 crc kubenswrapper[4848]: I1204 14:15:23.836987 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerStarted","Data":"11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a"} Dec 04 14:15:24 crc kubenswrapper[4848]: I1204 14:15:24.416225 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="576a9a23-e104-4d55-8100-ee23628af7cf" path="/var/lib/kubelet/pods/576a9a23-e104-4d55-8100-ee23628af7cf/volumes" Dec 04 14:15:24 crc kubenswrapper[4848]: I1204 14:15:24.852038 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerStarted","Data":"52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b"} Dec 04 14:15:24 crc kubenswrapper[4848]: I1204 14:15:24.852377 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:15:24 crc kubenswrapper[4848]: I1204 14:15:24.882707 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.684165767 podStartE2EDuration="4.882687116s" podCreationTimestamp="2025-12-04 14:15:20 +0000 UTC" firstStartedPulling="2025-12-04 14:15:21.266934929 +0000 UTC m=+1625.209431457" lastFinishedPulling="2025-12-04 14:15:24.465456278 +0000 UTC m=+1628.407952806" observedRunningTime="2025-12-04 14:15:24.882605574 +0000 UTC m=+1628.825102102" watchObservedRunningTime="2025-12-04 14:15:24.882687116 +0000 UTC m=+1628.825183644" Dec 04 14:15:27 crc kubenswrapper[4848]: I1204 14:15:27.887917 4848 generic.go:334] "Generic (PLEG): container finished" podID="1607ed31-5235-4f0c-a8e9-90f428f1da21" containerID="1a0303464165f9c664f35f880812155bdf885db86f9276f9bb584f140fe87c8a" exitCode=0 Dec 04 14:15:27 crc kubenswrapper[4848]: I1204 14:15:27.887981 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-msdsj" event={"ID":"1607ed31-5235-4f0c-a8e9-90f428f1da21","Type":"ContainerDied","Data":"1a0303464165f9c664f35f880812155bdf885db86f9276f9bb584f140fe87c8a"} Dec 04 14:15:28 crc kubenswrapper[4848]: I1204 14:15:28.394333 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:15:28 crc kubenswrapper[4848]: E1204 14:15:28.394829 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.377209 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.485808 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.485856 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.505763 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-scripts\") pod \"1607ed31-5235-4f0c-a8e9-90f428f1da21\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.506223 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwmgr\" (UniqueName: \"kubernetes.io/projected/1607ed31-5235-4f0c-a8e9-90f428f1da21-kube-api-access-vwmgr\") pod \"1607ed31-5235-4f0c-a8e9-90f428f1da21\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.506366 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-config-data\") pod \"1607ed31-5235-4f0c-a8e9-90f428f1da21\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.506428 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-combined-ca-bundle\") pod \"1607ed31-5235-4f0c-a8e9-90f428f1da21\" (UID: \"1607ed31-5235-4f0c-a8e9-90f428f1da21\") " Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.512146 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1607ed31-5235-4f0c-a8e9-90f428f1da21-kube-api-access-vwmgr" (OuterVolumeSpecName: "kube-api-access-vwmgr") pod "1607ed31-5235-4f0c-a8e9-90f428f1da21" (UID: "1607ed31-5235-4f0c-a8e9-90f428f1da21"). InnerVolumeSpecName "kube-api-access-vwmgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.512467 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-scripts" (OuterVolumeSpecName: "scripts") pod "1607ed31-5235-4f0c-a8e9-90f428f1da21" (UID: "1607ed31-5235-4f0c-a8e9-90f428f1da21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.543785 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-config-data" (OuterVolumeSpecName: "config-data") pod "1607ed31-5235-4f0c-a8e9-90f428f1da21" (UID: "1607ed31-5235-4f0c-a8e9-90f428f1da21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.544428 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1607ed31-5235-4f0c-a8e9-90f428f1da21" (UID: "1607ed31-5235-4f0c-a8e9-90f428f1da21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.610530 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwmgr\" (UniqueName: \"kubernetes.io/projected/1607ed31-5235-4f0c-a8e9-90f428f1da21-kube-api-access-vwmgr\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.610866 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.610881 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.610893 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1607ed31-5235-4f0c-a8e9-90f428f1da21-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.839049 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.840995 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.848215 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.911657 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-msdsj" event={"ID":"1607ed31-5235-4f0c-a8e9-90f428f1da21","Type":"ContainerDied","Data":"84b377bc08d8914b37119cef1e3a414e8e199dd16c789f147834dfce68f59c85"} Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.911690 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-msdsj" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.911693 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84b377bc08d8914b37119cef1e3a414e8e199dd16c789f147834dfce68f59c85" Dec 04 14:15:29 crc kubenswrapper[4848]: I1204 14:15:29.918900 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.103120 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.103497 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-log" containerID="cri-o://a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a" gracePeriod=30 Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.103550 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-api" containerID="cri-o://1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377" gracePeriod=30 Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.125814 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.126192 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f7c7d746-2b41-4319-9342-bdca97131d23" containerName="nova-scheduler-scheduler" containerID="cri-o://362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665" gracePeriod=30 Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.132874 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.251:8774/\": EOF" Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.135226 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.251:8774/\": EOF" Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.139447 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.930077 4848 generic.go:334] "Generic (PLEG): container finished" podID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerID="a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a" exitCode=143 Dec 04 14:15:30 crc kubenswrapper[4848]: I1204 14:15:30.930187 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3fbac58-059d-4bab-b329-aa13c7a4552c","Type":"ContainerDied","Data":"a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a"} Dec 04 14:15:31 crc kubenswrapper[4848]: I1204 14:15:31.939348 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-log" containerID="cri-o://592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133" gracePeriod=30 Dec 04 14:15:31 crc kubenswrapper[4848]: I1204 14:15:31.939428 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-metadata" containerID="cri-o://bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9" gracePeriod=30 Dec 04 14:15:32 crc kubenswrapper[4848]: I1204 14:15:32.954014 4848 generic.go:334] "Generic (PLEG): container finished" podID="1846d8fa-57ed-495f-aca9-87369c79279b" containerID="592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133" exitCode=143 Dec 04 14:15:32 crc kubenswrapper[4848]: I1204 14:15:32.954072 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1846d8fa-57ed-495f-aca9-87369c79279b","Type":"ContainerDied","Data":"592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133"} Dec 04 14:15:34 crc kubenswrapper[4848]: E1204 14:15:34.466103 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665 is running failed: container process not found" containerID="362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:15:34 crc kubenswrapper[4848]: E1204 14:15:34.466730 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665 is running failed: container process not found" containerID="362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:15:34 crc kubenswrapper[4848]: E1204 14:15:34.466985 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665 is running failed: container process not found" containerID="362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:15:34 crc kubenswrapper[4848]: E1204 14:15:34.467016 4848 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f7c7d746-2b41-4319-9342-bdca97131d23" containerName="nova-scheduler-scheduler" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.500506 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.630830 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-config-data\") pod \"f7c7d746-2b41-4319-9342-bdca97131d23\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.630887 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtjv4\" (UniqueName: \"kubernetes.io/projected/f7c7d746-2b41-4319-9342-bdca97131d23-kube-api-access-gtjv4\") pod \"f7c7d746-2b41-4319-9342-bdca97131d23\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.631275 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-combined-ca-bundle\") pod \"f7c7d746-2b41-4319-9342-bdca97131d23\" (UID: \"f7c7d746-2b41-4319-9342-bdca97131d23\") " Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.636350 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c7d746-2b41-4319-9342-bdca97131d23-kube-api-access-gtjv4" (OuterVolumeSpecName: "kube-api-access-gtjv4") pod "f7c7d746-2b41-4319-9342-bdca97131d23" (UID: "f7c7d746-2b41-4319-9342-bdca97131d23"). InnerVolumeSpecName "kube-api-access-gtjv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.663235 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7c7d746-2b41-4319-9342-bdca97131d23" (UID: "f7c7d746-2b41-4319-9342-bdca97131d23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.671129 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-config-data" (OuterVolumeSpecName: "config-data") pod "f7c7d746-2b41-4319-9342-bdca97131d23" (UID: "f7c7d746-2b41-4319-9342-bdca97131d23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.733775 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.733807 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c7d746-2b41-4319-9342-bdca97131d23-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.733816 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtjv4\" (UniqueName: \"kubernetes.io/projected/f7c7d746-2b41-4319-9342-bdca97131d23-kube-api-access-gtjv4\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.981686 4848 generic.go:334] "Generic (PLEG): container finished" podID="f7c7d746-2b41-4319-9342-bdca97131d23" containerID="362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665" exitCode=0 Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.981750 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7c7d746-2b41-4319-9342-bdca97131d23","Type":"ContainerDied","Data":"362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665"} Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.981791 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7c7d746-2b41-4319-9342-bdca97131d23","Type":"ContainerDied","Data":"00c26ea5e51202456c01c045d6e5021db0eb4121dd94459c01153924bd08e78e"} Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.981792 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:15:34 crc kubenswrapper[4848]: I1204 14:15:34.981815 4848 scope.go:117] "RemoveContainer" containerID="362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.050838 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.057298 4848 scope.go:117] "RemoveContainer" containerID="362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665" Dec 04 14:15:35 crc kubenswrapper[4848]: E1204 14:15:35.058558 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665\": container with ID starting with 362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665 not found: ID does not exist" containerID="362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.058616 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665"} err="failed to get container status \"362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665\": rpc error: code = NotFound desc = could not find container \"362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665\": container with ID starting with 362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665 not found: ID does not exist" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.079966 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.094244 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": read tcp 10.217.0.2:55446->10.217.0.248:8775: read: connection reset by peer" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.094249 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": read tcp 10.217.0.2:55448->10.217.0.248:8775: read: connection reset by peer" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.104247 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:15:35 crc kubenswrapper[4848]: E1204 14:15:35.104843 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="576a9a23-e104-4d55-8100-ee23628af7cf" containerName="init" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.104870 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="576a9a23-e104-4d55-8100-ee23628af7cf" containerName="init" Dec 04 14:15:35 crc kubenswrapper[4848]: E1204 14:15:35.104890 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c7d746-2b41-4319-9342-bdca97131d23" containerName="nova-scheduler-scheduler" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.104898 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c7d746-2b41-4319-9342-bdca97131d23" containerName="nova-scheduler-scheduler" Dec 04 14:15:35 crc kubenswrapper[4848]: E1204 14:15:35.104914 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="576a9a23-e104-4d55-8100-ee23628af7cf" containerName="dnsmasq-dns" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.104923 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="576a9a23-e104-4d55-8100-ee23628af7cf" containerName="dnsmasq-dns" Dec 04 14:15:35 crc kubenswrapper[4848]: E1204 14:15:35.104965 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1607ed31-5235-4f0c-a8e9-90f428f1da21" containerName="nova-manage" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.104974 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="1607ed31-5235-4f0c-a8e9-90f428f1da21" containerName="nova-manage" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.105296 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="576a9a23-e104-4d55-8100-ee23628af7cf" containerName="dnsmasq-dns" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.105318 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="1607ed31-5235-4f0c-a8e9-90f428f1da21" containerName="nova-manage" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.105331 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c7d746-2b41-4319-9342-bdca97131d23" containerName="nova-scheduler-scheduler" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.106413 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.109667 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.123852 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.260667 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxjb9\" (UniqueName: \"kubernetes.io/projected/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-kube-api-access-hxjb9\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.261060 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.261156 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-config-data\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.363539 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-config-data\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.363672 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxjb9\" (UniqueName: \"kubernetes.io/projected/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-kube-api-access-hxjb9\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.363751 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.368634 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-config-data\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.374736 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.383107 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxjb9\" (UniqueName: \"kubernetes.io/projected/7fa9a476-cd4b-4258-ba07-0ef2888fa4e3-kube-api-access-hxjb9\") pod \"nova-scheduler-0\" (UID: \"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3\") " pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.550175 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.702716 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.774703 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1846d8fa-57ed-495f-aca9-87369c79279b-logs\") pod \"1846d8fa-57ed-495f-aca9-87369c79279b\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.774757 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmx8r\" (UniqueName: \"kubernetes.io/projected/1846d8fa-57ed-495f-aca9-87369c79279b-kube-api-access-fmx8r\") pod \"1846d8fa-57ed-495f-aca9-87369c79279b\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.775029 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-combined-ca-bundle\") pod \"1846d8fa-57ed-495f-aca9-87369c79279b\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.775074 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-nova-metadata-tls-certs\") pod \"1846d8fa-57ed-495f-aca9-87369c79279b\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.775105 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-config-data\") pod \"1846d8fa-57ed-495f-aca9-87369c79279b\" (UID: \"1846d8fa-57ed-495f-aca9-87369c79279b\") " Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.775312 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1846d8fa-57ed-495f-aca9-87369c79279b-logs" (OuterVolumeSpecName: "logs") pod "1846d8fa-57ed-495f-aca9-87369c79279b" (UID: "1846d8fa-57ed-495f-aca9-87369c79279b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.775853 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1846d8fa-57ed-495f-aca9-87369c79279b-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.780354 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1846d8fa-57ed-495f-aca9-87369c79279b-kube-api-access-fmx8r" (OuterVolumeSpecName: "kube-api-access-fmx8r") pod "1846d8fa-57ed-495f-aca9-87369c79279b" (UID: "1846d8fa-57ed-495f-aca9-87369c79279b"). InnerVolumeSpecName "kube-api-access-fmx8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.815551 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1846d8fa-57ed-495f-aca9-87369c79279b" (UID: "1846d8fa-57ed-495f-aca9-87369c79279b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.818078 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-config-data" (OuterVolumeSpecName: "config-data") pod "1846d8fa-57ed-495f-aca9-87369c79279b" (UID: "1846d8fa-57ed-495f-aca9-87369c79279b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.861618 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1846d8fa-57ed-495f-aca9-87369c79279b" (UID: "1846d8fa-57ed-495f-aca9-87369c79279b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.877429 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.877461 4848 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.877474 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1846d8fa-57ed-495f-aca9-87369c79279b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:35 crc kubenswrapper[4848]: I1204 14:15:35.877484 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmx8r\" (UniqueName: \"kubernetes.io/projected/1846d8fa-57ed-495f-aca9-87369c79279b-kube-api-access-fmx8r\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.004766 4848 generic.go:334] "Generic (PLEG): container finished" podID="1846d8fa-57ed-495f-aca9-87369c79279b" containerID="bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9" exitCode=0 Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.005079 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1846d8fa-57ed-495f-aca9-87369c79279b","Type":"ContainerDied","Data":"bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9"} Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.005107 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1846d8fa-57ed-495f-aca9-87369c79279b","Type":"ContainerDied","Data":"6fc76e2463c5817b3dfb2c467b555d334d26697d47334ff568899049694b7372"} Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.005123 4848 scope.go:117] "RemoveContainer" containerID="bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.005230 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.026912 4848 generic.go:334] "Generic (PLEG): container finished" podID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerID="1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377" exitCode=0 Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.026999 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3fbac58-059d-4bab-b329-aa13c7a4552c","Type":"ContainerDied","Data":"1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377"} Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.053702 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.074019 4848 scope.go:117] "RemoveContainer" containerID="592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.082685 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.105090 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.106207 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.118759 4848 scope.go:117] "RemoveContainer" containerID="bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9" Dec 04 14:15:36 crc kubenswrapper[4848]: E1204 14:15:36.119314 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9\": container with ID starting with bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9 not found: ID does not exist" containerID="bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.119352 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9"} err="failed to get container status \"bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9\": rpc error: code = NotFound desc = could not find container \"bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9\": container with ID starting with bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9 not found: ID does not exist" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.119406 4848 scope.go:117] "RemoveContainer" containerID="592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133" Dec 04 14:15:36 crc kubenswrapper[4848]: E1204 14:15:36.120207 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133\": container with ID starting with 592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133 not found: ID does not exist" containerID="592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.120236 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133"} err="failed to get container status \"592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133\": rpc error: code = NotFound desc = could not find container \"592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133\": container with ID starting with 592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133 not found: ID does not exist" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.139033 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:36 crc kubenswrapper[4848]: E1204 14:15:36.139665 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-metadata" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.139687 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-metadata" Dec 04 14:15:36 crc kubenswrapper[4848]: E1204 14:15:36.139715 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-log" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.139722 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-log" Dec 04 14:15:36 crc kubenswrapper[4848]: E1204 14:15:36.139736 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-api" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.139744 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-api" Dec 04 14:15:36 crc kubenswrapper[4848]: E1204 14:15:36.139789 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-log" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.139796 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-log" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.140077 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-log" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.140102 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-log" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.140134 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" containerName="nova-metadata-metadata" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.140149 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" containerName="nova-api-api" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.141677 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.145966 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.146615 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.153868 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.184982 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9487\" (UniqueName: \"kubernetes.io/projected/b3fbac58-059d-4bab-b329-aa13c7a4552c-kube-api-access-n9487\") pod \"b3fbac58-059d-4bab-b329-aa13c7a4552c\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.185043 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-internal-tls-certs\") pod \"b3fbac58-059d-4bab-b329-aa13c7a4552c\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.185136 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-combined-ca-bundle\") pod \"b3fbac58-059d-4bab-b329-aa13c7a4552c\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.185164 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3fbac58-059d-4bab-b329-aa13c7a4552c-logs\") pod \"b3fbac58-059d-4bab-b329-aa13c7a4552c\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.185190 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-public-tls-certs\") pod \"b3fbac58-059d-4bab-b329-aa13c7a4552c\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.185372 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-config-data\") pod \"b3fbac58-059d-4bab-b329-aa13c7a4552c\" (UID: \"b3fbac58-059d-4bab-b329-aa13c7a4552c\") " Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.186128 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3fbac58-059d-4bab-b329-aa13c7a4552c-logs" (OuterVolumeSpecName: "logs") pod "b3fbac58-059d-4bab-b329-aa13c7a4552c" (UID: "b3fbac58-059d-4bab-b329-aa13c7a4552c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.189818 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3fbac58-059d-4bab-b329-aa13c7a4552c-kube-api-access-n9487" (OuterVolumeSpecName: "kube-api-access-n9487") pod "b3fbac58-059d-4bab-b329-aa13c7a4552c" (UID: "b3fbac58-059d-4bab-b329-aa13c7a4552c"). InnerVolumeSpecName "kube-api-access-n9487". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.222363 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3fbac58-059d-4bab-b329-aa13c7a4552c" (UID: "b3fbac58-059d-4bab-b329-aa13c7a4552c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.224767 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-config-data" (OuterVolumeSpecName: "config-data") pod "b3fbac58-059d-4bab-b329-aa13c7a4552c" (UID: "b3fbac58-059d-4bab-b329-aa13c7a4552c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.252586 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b3fbac58-059d-4bab-b329-aa13c7a4552c" (UID: "b3fbac58-059d-4bab-b329-aa13c7a4552c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.258254 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b3fbac58-059d-4bab-b329-aa13c7a4552c" (UID: "b3fbac58-059d-4bab-b329-aa13c7a4552c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.288484 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whqwb\" (UniqueName: \"kubernetes.io/projected/42fe17d3-9143-40e0-ab66-074885ccd028-kube-api-access-whqwb\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.288535 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-config-data\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.288885 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.289135 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.289203 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42fe17d3-9143-40e0-ab66-074885ccd028-logs\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.289371 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9487\" (UniqueName: \"kubernetes.io/projected/b3fbac58-059d-4bab-b329-aa13c7a4552c-kube-api-access-n9487\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.289390 4848 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.289401 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.289410 4848 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3fbac58-059d-4bab-b329-aa13c7a4552c-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.289419 4848 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.289428 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3fbac58-059d-4bab-b329-aa13c7a4552c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.390872 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whqwb\" (UniqueName: \"kubernetes.io/projected/42fe17d3-9143-40e0-ab66-074885ccd028-kube-api-access-whqwb\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.390910 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-config-data\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.391039 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.391102 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.391138 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42fe17d3-9143-40e0-ab66-074885ccd028-logs\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.391572 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42fe17d3-9143-40e0-ab66-074885ccd028-logs\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.394594 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-config-data\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.394694 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.401931 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/42fe17d3-9143-40e0-ab66-074885ccd028-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.409183 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whqwb\" (UniqueName: \"kubernetes.io/projected/42fe17d3-9143-40e0-ab66-074885ccd028-kube-api-access-whqwb\") pod \"nova-metadata-0\" (UID: \"42fe17d3-9143-40e0-ab66-074885ccd028\") " pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.428387 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1846d8fa-57ed-495f-aca9-87369c79279b" path="/var/lib/kubelet/pods/1846d8fa-57ed-495f-aca9-87369c79279b/volumes" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.429160 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c7d746-2b41-4319-9342-bdca97131d23" path="/var/lib/kubelet/pods/f7c7d746-2b41-4319-9342-bdca97131d23/volumes" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.473611 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:15:36 crc kubenswrapper[4848]: I1204 14:15:36.893058 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.049537 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b3fbac58-059d-4bab-b329-aa13c7a4552c","Type":"ContainerDied","Data":"4ff1adfa7e2d3615dbba8d77021be244a62dc0281dfbccac220fe4748a434338"} Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.049606 4848 scope.go:117] "RemoveContainer" containerID="1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.049774 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.059648 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3","Type":"ContainerStarted","Data":"565f37621b330202212ab64183c7026aafd8ed643037fdbedab9cf21ac30a841"} Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.059684 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7fa9a476-cd4b-4258-ba07-0ef2888fa4e3","Type":"ContainerStarted","Data":"b19f2a8cabee78616e59e3d2856c0bd9ac1f47c0a38c1056648a89759ac9aeef"} Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.063993 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42fe17d3-9143-40e0-ab66-074885ccd028","Type":"ContainerStarted","Data":"4b456fee4041f7e8e72c20e1b8bf288d807778c4dfaa0527ce0e15c6fcdf8785"} Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.086323 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.101998 4848 scope.go:117] "RemoveContainer" containerID="a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.105323 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.117031 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.1170091 podStartE2EDuration="2.1170091s" podCreationTimestamp="2025-12-04 14:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:37.088166 +0000 UTC m=+1641.030662528" watchObservedRunningTime="2025-12-04 14:15:37.1170091 +0000 UTC m=+1641.059505628" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.158744 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.172485 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.191067 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.191402 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.192085 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.206463 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.319922 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.320350 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-public-tls-certs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.320518 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-config-data\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.320614 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.320662 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9781260-81d8-4950-8388-00ae237a6f7f-logs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.320722 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx2zb\" (UniqueName: \"kubernetes.io/projected/d9781260-81d8-4950-8388-00ae237a6f7f-kube-api-access-tx2zb\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.423087 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx2zb\" (UniqueName: \"kubernetes.io/projected/d9781260-81d8-4950-8388-00ae237a6f7f-kube-api-access-tx2zb\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.423459 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.423617 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-public-tls-certs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.423817 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-config-data\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.424013 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.424555 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9781260-81d8-4950-8388-00ae237a6f7f-logs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.425166 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9781260-81d8-4950-8388-00ae237a6f7f-logs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.427811 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.428373 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-config-data\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.428539 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-public-tls-certs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.439541 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9781260-81d8-4950-8388-00ae237a6f7f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.440864 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx2zb\" (UniqueName: \"kubernetes.io/projected/d9781260-81d8-4950-8388-00ae237a6f7f-kube-api-access-tx2zb\") pod \"nova-api-0\" (UID: \"d9781260-81d8-4950-8388-00ae237a6f7f\") " pod="openstack/nova-api-0" Dec 04 14:15:37 crc kubenswrapper[4848]: I1204 14:15:37.580871 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:15:38 crc kubenswrapper[4848]: E1204 14:15:38.055787 4848 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/dea2b3f1579678a8b0bafc288689d36bc89e6cb94644e4f77d0b2ff51317e2f8/diff" to get inode usage: stat /var/lib/containers/storage/overlay/dea2b3f1579678a8b0bafc288689d36bc89e6cb94644e4f77d0b2ff51317e2f8/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-9b86998b5-rx28f_576a9a23-e104-4d55-8100-ee23628af7cf/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-9b86998b5-rx28f_576a9a23-e104-4d55-8100-ee23628af7cf/dnsmasq-dns/0.log: no such file or directory Dec 04 14:15:38 crc kubenswrapper[4848]: I1204 14:15:38.074366 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:15:38 crc kubenswrapper[4848]: I1204 14:15:38.075136 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42fe17d3-9143-40e0-ab66-074885ccd028","Type":"ContainerStarted","Data":"f4b3b4670c4fbcb8c49edd27d3e7ac07a84e5b94584b3468d26cc50a7bfe0658"} Dec 04 14:15:38 crc kubenswrapper[4848]: I1204 14:15:38.075181 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42fe17d3-9143-40e0-ab66-074885ccd028","Type":"ContainerStarted","Data":"9f0926f732d0a86c9041257d150624fa0b8a316c21fd2bc3d122cbc659135e65"} Dec 04 14:15:38 crc kubenswrapper[4848]: I1204 14:15:38.094477 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.094460076 podStartE2EDuration="2.094460076s" podCreationTimestamp="2025-12-04 14:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:38.093916053 +0000 UTC m=+1642.036412581" watchObservedRunningTime="2025-12-04 14:15:38.094460076 +0000 UTC m=+1642.036956604" Dec 04 14:15:38 crc kubenswrapper[4848]: I1204 14:15:38.406327 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3fbac58-059d-4bab-b329-aa13c7a4552c" path="/var/lib/kubelet/pods/b3fbac58-059d-4bab-b329-aa13c7a4552c/volumes" Dec 04 14:15:39 crc kubenswrapper[4848]: I1204 14:15:39.089743 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9781260-81d8-4950-8388-00ae237a6f7f","Type":"ContainerStarted","Data":"7e8adec590893862936e56c1403773ab5c997e2af4c8984400740af425abc1f5"} Dec 04 14:15:39 crc kubenswrapper[4848]: I1204 14:15:39.089799 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9781260-81d8-4950-8388-00ae237a6f7f","Type":"ContainerStarted","Data":"ed339368b91baeb27a442e1bda77bb6662754d7323daeb5ab92c6b8ad5eb03c9"} Dec 04 14:15:39 crc kubenswrapper[4848]: I1204 14:15:39.089811 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d9781260-81d8-4950-8388-00ae237a6f7f","Type":"ContainerStarted","Data":"d420773adb601a95bf61946a0ba5409939929ca4366a677c4983f9bfd7b7be6e"} Dec 04 14:15:39 crc kubenswrapper[4848]: I1204 14:15:39.111205 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.111186076 podStartE2EDuration="2.111186076s" podCreationTimestamp="2025-12-04 14:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:15:39.105672023 +0000 UTC m=+1643.048168551" watchObservedRunningTime="2025-12-04 14:15:39.111186076 +0000 UTC m=+1643.053682604" Dec 04 14:15:39 crc kubenswrapper[4848]: I1204 14:15:39.393135 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:15:39 crc kubenswrapper[4848]: E1204 14:15:39.393885 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:15:40 crc kubenswrapper[4848]: W1204 14:15:40.450379 4848 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1607ed31_5235_4f0c_a8e9_90f428f1da21.slice/crio-84b377bc08d8914b37119cef1e3a414e8e199dd16c789f147834dfce68f59c85": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1607ed31_5235_4f0c_a8e9_90f428f1da21.slice/crio-84b377bc08d8914b37119cef1e3a414e8e199dd16c789f147834dfce68f59c85: no such file or directory Dec 04 14:15:40 crc kubenswrapper[4848]: W1204 14:15:40.450454 4848 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1607ed31_5235_4f0c_a8e9_90f428f1da21.slice/crio-conmon-1a0303464165f9c664f35f880812155bdf885db86f9276f9bb584f140fe87c8a.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1607ed31_5235_4f0c_a8e9_90f428f1da21.slice/crio-conmon-1a0303464165f9c664f35f880812155bdf885db86f9276f9bb584f140fe87c8a.scope: no such file or directory Dec 04 14:15:40 crc kubenswrapper[4848]: W1204 14:15:40.450474 4848 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1607ed31_5235_4f0c_a8e9_90f428f1da21.slice/crio-1a0303464165f9c664f35f880812155bdf885db86f9276f9bb584f140fe87c8a.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1607ed31_5235_4f0c_a8e9_90f428f1da21.slice/crio-1a0303464165f9c664f35f880812155bdf885db86f9276f9bb584f140fe87c8a.scope: no such file or directory Dec 04 14:15:40 crc kubenswrapper[4848]: W1204 14:15:40.457968 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3fbac58_059d_4bab_b329_aa13c7a4552c.slice/crio-a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a.scope WatchSource:0}: Error finding container a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a: Status 404 returned error can't find the container with id a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a Dec 04 14:15:40 crc kubenswrapper[4848]: W1204 14:15:40.458685 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3fbac58_059d_4bab_b329_aa13c7a4552c.slice/crio-1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377.scope WatchSource:0}: Error finding container 1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377: Status 404 returned error can't find the container with id 1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377 Dec 04 14:15:40 crc kubenswrapper[4848]: I1204 14:15:40.550247 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 14:15:40 crc kubenswrapper[4848]: E1204 14:15:40.587667 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee226d6d_c7ec_4125_bdf4_c96c3ec0dfb3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d74c53_ba76_41a2_a258_a0e45f3dc2ef.slice/crio-conmon-6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice/crio-conmon-362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-conmon-bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice/crio-conmon-18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee226d6d_c7ec_4125_bdf4_c96c3ec0dfb3.slice/crio-conmon-63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice/crio-18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d74c53_ba76_41a2_a258_a0e45f3dc2ef.slice/crio-1c38bcadac24308d808b1415936baee05edc298fffe0518b79e884694e444bca\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice/crio-547e7e3cd3211e334c27881428a32645df5ac620fe0df7250b0de191e7bfe063\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d74c53_ba76_41a2_a258_a0e45f3dc2ef.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ba3e54_9bc1_49f4_ad14_195c52649b8c.slice/crio-conmon-cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d74c53_ba76_41a2_a258_a0e45f3dc2ef.slice/crio-6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-conmon-592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-6fc76e2463c5817b3dfb2c467b555d334d26697d47334ff568899049694b7372\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ba3e54_9bc1_49f4_ad14_195c52649b8c.slice/crio-cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice/crio-362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee226d6d_c7ec_4125_bdf4_c96c3ec0dfb3.slice/crio-0d8ed6cbcdb23009b16b99575fd8ee715bf4dd5817f9f64750dc8fc95fc97d98\": RecentStats: unable to find data in memory cache]" Dec 04 14:15:40 crc kubenswrapper[4848]: E1204 14:15:40.587933 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee226d6d_c7ec_4125_bdf4_c96c3ec0dfb3.slice/crio-63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d74c53_ba76_41a2_a258_a0e45f3dc2ef.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-conmon-bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ba3e54_9bc1_49f4_ad14_195c52649b8c.slice/crio-conmon-cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d74c53_ba76_41a2_a258_a0e45f3dc2ef.slice/crio-conmon-6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee226d6d_c7ec_4125_bdf4_c96c3ec0dfb3.slice/crio-0d8ed6cbcdb23009b16b99575fd8ee715bf4dd5817f9f64750dc8fc95fc97d98\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ba3e54_9bc1_49f4_ad14_195c52649b8c.slice/crio-cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice/crio-547e7e3cd3211e334c27881428a32645df5ac620fe0df7250b0de191e7bfe063\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice/crio-conmon-18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee226d6d_c7ec_4125_bdf4_c96c3ec0dfb3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-6fc76e2463c5817b3dfb2c467b555d334d26697d47334ff568899049694b7372\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee226d6d_c7ec_4125_bdf4_c96c3ec0dfb3.slice/crio-conmon-63b292b9bce73d847dce2aec96e62e70029675b3a1ee9860a898544a6e2c9694.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d74c53_ba76_41a2_a258_a0e45f3dc2ef.slice/crio-1c38bcadac24308d808b1415936baee05edc298fffe0518b79e884694e444bca\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d74c53_ba76_41a2_a258_a0e45f3dc2ef.slice/crio-6f00063f3049cffd653bbae6a0c525450dc82dc980a0e42a3fc0c657921f1f0a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice/crio-18a279f0068eac3d524c7acc693cbfa32ead378f50ddc1c863c5d373cfd2ba13.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice/crio-362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:15:40 crc kubenswrapper[4848]: E1204 14:15:40.590967 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3fbac58_059d_4bab_b329_aa13c7a4552c.slice/crio-conmon-1b4b5753028ad52db46718e4b391b5e40a51d59e7ca6902e5c7d85fac1aae377.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice/crio-362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-6fc76e2463c5817b3dfb2c467b555d334d26697d47334ff568899049694b7372\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-conmon-bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3fbac58_059d_4bab_b329_aa13c7a4552c.slice/crio-4ff1adfa7e2d3615dbba8d77021be244a62dc0281dfbccac220fe4748a434338\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice/crio-00c26ea5e51202456c01c045d6e5021db0eb4121dd94459c01153924bd08e78e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-conmon-592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-bc2f99a24d61f6fb0de1a38736cd49af96534fa5de053c42b19381b9b38ea0c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3fbac58_059d_4bab_b329_aa13c7a4552c.slice/crio-conmon-a3a661f3599e2994d8401d9336acb834ee9eed64450625bf9c636dc14a24370a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1846d8fa_57ed_495f_aca9_87369c79279b.slice/crio-592182c186cf87988216ffaff5d64803e761ddd632043df98593be28839db133.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ba3e54_9bc1_49f4_ad14_195c52649b8c.slice/crio-cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576a9a23_e104_4d55_8100_ee23628af7cf.slice/crio-547e7e3cd3211e334c27881428a32645df5ac620fe0df7250b0de191e7bfe063\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1607ed31_5235_4f0c_a8e9_90f428f1da21.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c7d746_2b41_4319_9342_bdca97131d23.slice/crio-conmon-362f226965d633196bdcad4520b32f1202ef9384a2d0b5ed24e6d2cbd2bc7665.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3fbac58_059d_4bab_b329_aa13c7a4552c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01ba3e54_9bc1_49f4_ad14_195c52649b8c.slice/crio-conmon-cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:15:41 crc kubenswrapper[4848]: I1204 14:15:41.224457 4848 generic.go:334] "Generic (PLEG): container finished" podID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerID="cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f" exitCode=137 Dec 04 14:15:41 crc kubenswrapper[4848]: I1204 14:15:41.224505 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerDied","Data":"cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f"} Dec 04 14:15:41 crc kubenswrapper[4848]: I1204 14:15:41.474280 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:15:41 crc kubenswrapper[4848]: I1204 14:15:41.474625 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:15:41 crc kubenswrapper[4848]: I1204 14:15:41.941642 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.046181 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-scripts\") pod \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.046404 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-combined-ca-bundle\") pod \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.046524 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-config-data\") pod \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.046639 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntkgq\" (UniqueName: \"kubernetes.io/projected/01ba3e54-9bc1-49f4-ad14-195c52649b8c-kube-api-access-ntkgq\") pod \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\" (UID: \"01ba3e54-9bc1-49f4-ad14-195c52649b8c\") " Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.053103 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-scripts" (OuterVolumeSpecName: "scripts") pod "01ba3e54-9bc1-49f4-ad14-195c52649b8c" (UID: "01ba3e54-9bc1-49f4-ad14-195c52649b8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.080155 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ba3e54-9bc1-49f4-ad14-195c52649b8c-kube-api-access-ntkgq" (OuterVolumeSpecName: "kube-api-access-ntkgq") pod "01ba3e54-9bc1-49f4-ad14-195c52649b8c" (UID: "01ba3e54-9bc1-49f4-ad14-195c52649b8c"). InnerVolumeSpecName "kube-api-access-ntkgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.149517 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.149548 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntkgq\" (UniqueName: \"kubernetes.io/projected/01ba3e54-9bc1-49f4-ad14-195c52649b8c-kube-api-access-ntkgq\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.212916 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01ba3e54-9bc1-49f4-ad14-195c52649b8c" (UID: "01ba3e54-9bc1-49f4-ad14-195c52649b8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.238103 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"01ba3e54-9bc1-49f4-ad14-195c52649b8c","Type":"ContainerDied","Data":"d3a13123ffbba2c50e0890a2581b196deb64757bca23ec5b0b177d455f35fc9e"} Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.238148 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.238167 4848 scope.go:117] "RemoveContainer" containerID="cc8e02df8d82e9c3f1566e9e2f5c090f90a1fa67c86dfd4227934638ad9f219f" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.240723 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-config-data" (OuterVolumeSpecName: "config-data") pod "01ba3e54-9bc1-49f4-ad14-195c52649b8c" (UID: "01ba3e54-9bc1-49f4-ad14-195c52649b8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.251539 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.251570 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ba3e54-9bc1-49f4-ad14-195c52649b8c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.261506 4848 scope.go:117] "RemoveContainer" containerID="c70373684745f5ec0a9f579de707b989e320e7b0328e827e16b27423d32f16f8" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.282042 4848 scope.go:117] "RemoveContainer" containerID="636cd099d4df528e25e0936f756ab1d111dc07ec30ca3db85abb7895f820c46d" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.306596 4848 scope.go:117] "RemoveContainer" containerID="a49f648dcbaf53078904342a3ca35d6f3568275890ed9e58c49a772698c70288" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.564476 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.581035 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.594985 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 04 14:15:42 crc kubenswrapper[4848]: E1204 14:15:42.595446 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-api" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.595462 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-api" Dec 04 14:15:42 crc kubenswrapper[4848]: E1204 14:15:42.595484 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-notifier" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.595490 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-notifier" Dec 04 14:15:42 crc kubenswrapper[4848]: E1204 14:15:42.595510 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-listener" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.595516 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-listener" Dec 04 14:15:42 crc kubenswrapper[4848]: E1204 14:15:42.595592 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-evaluator" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.595599 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-evaluator" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.595883 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-api" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.595913 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-evaluator" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.595932 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-listener" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.595972 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" containerName="aodh-notifier" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.598393 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.602934 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-d7qhp" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.603498 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.603922 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.605800 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.608059 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.615332 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.664907 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-config-data\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.664990 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-public-tls-certs\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.665098 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-internal-tls-certs\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.665235 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.665326 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-scripts\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.665386 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpbb6\" (UniqueName: \"kubernetes.io/projected/a570d885-c380-4d05-af91-890fa9eb893b-kube-api-access-dpbb6\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.768487 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-config-data\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.768569 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-public-tls-certs\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.768679 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-internal-tls-certs\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.768809 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.768934 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-scripts\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.769040 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpbb6\" (UniqueName: \"kubernetes.io/projected/a570d885-c380-4d05-af91-890fa9eb893b-kube-api-access-dpbb6\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.774882 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-scripts\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.775314 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-config-data\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.775390 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-internal-tls-certs\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.775705 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.775855 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-public-tls-certs\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.791823 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpbb6\" (UniqueName: \"kubernetes.io/projected/a570d885-c380-4d05-af91-890fa9eb893b-kube-api-access-dpbb6\") pod \"aodh-0\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " pod="openstack/aodh-0" Dec 04 14:15:42 crc kubenswrapper[4848]: I1204 14:15:42.915061 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:15:43 crc kubenswrapper[4848]: I1204 14:15:43.392035 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 14:15:44 crc kubenswrapper[4848]: I1204 14:15:44.275554 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerStarted","Data":"8d7254c20441ec1adb4671df1a7577740de25dea51bc7a57bfab8e723a317206"} Dec 04 14:15:44 crc kubenswrapper[4848]: I1204 14:15:44.276030 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerStarted","Data":"e5ac7a15469ee04ec9aabe605eaaaf71be05f2b012f86e0ffe3051065bad14ac"} Dec 04 14:15:44 crc kubenswrapper[4848]: I1204 14:15:44.409244 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ba3e54-9bc1-49f4-ad14-195c52649b8c" path="/var/lib/kubelet/pods/01ba3e54-9bc1-49f4-ad14-195c52649b8c/volumes" Dec 04 14:15:45 crc kubenswrapper[4848]: I1204 14:15:45.287909 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerStarted","Data":"7ef085a9fae995c05f374d9e2f2c080fc23a75bd76296891538d2733e97fbe2d"} Dec 04 14:15:45 crc kubenswrapper[4848]: I1204 14:15:45.550470 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 14:15:45 crc kubenswrapper[4848]: I1204 14:15:45.581693 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 14:15:46 crc kubenswrapper[4848]: I1204 14:15:46.357841 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 14:15:46 crc kubenswrapper[4848]: I1204 14:15:46.474650 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:15:46 crc kubenswrapper[4848]: I1204 14:15:46.474704 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:15:47 crc kubenswrapper[4848]: I1204 14:15:47.316181 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerStarted","Data":"f70f58c1a5d10fa2785184800d1b2560c76a9b5c4f17647a20f542ef2845c975"} Dec 04 14:15:47 crc kubenswrapper[4848]: I1204 14:15:47.488267 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="42fe17d3-9143-40e0-ab66-074885ccd028" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.255:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:15:47 crc kubenswrapper[4848]: I1204 14:15:47.488310 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="42fe17d3-9143-40e0-ab66-074885ccd028" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.255:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:15:47 crc kubenswrapper[4848]: I1204 14:15:47.582068 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:15:47 crc kubenswrapper[4848]: I1204 14:15:47.582117 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:15:48 crc kubenswrapper[4848]: I1204 14:15:48.346642 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerStarted","Data":"16e37ee5614acb0246bb4311be23a7b27c15fbba434517217cda5ab58abec44f"} Dec 04 14:15:48 crc kubenswrapper[4848]: I1204 14:15:48.373863 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.576341697 podStartE2EDuration="6.373840156s" podCreationTimestamp="2025-12-04 14:15:42 +0000 UTC" firstStartedPulling="2025-12-04 14:15:43.394465208 +0000 UTC m=+1647.336961736" lastFinishedPulling="2025-12-04 14:15:47.191963667 +0000 UTC m=+1651.134460195" observedRunningTime="2025-12-04 14:15:48.366213401 +0000 UTC m=+1652.308709929" watchObservedRunningTime="2025-12-04 14:15:48.373840156 +0000 UTC m=+1652.316336694" Dec 04 14:15:48 crc kubenswrapper[4848]: I1204 14:15:48.611410 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d9781260-81d8-4950-8388-00ae237a6f7f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.0:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:15:48 crc kubenswrapper[4848]: I1204 14:15:48.611677 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d9781260-81d8-4950-8388-00ae237a6f7f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.0:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:15:50 crc kubenswrapper[4848]: I1204 14:15:50.665021 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 14:15:54 crc kubenswrapper[4848]: I1204 14:15:54.394688 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:15:54 crc kubenswrapper[4848]: E1204 14:15:54.395749 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:15:54 crc kubenswrapper[4848]: I1204 14:15:54.552614 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:15:54 crc kubenswrapper[4848]: I1204 14:15:54.553116 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d644fdb6-4d43-4e46-9537-f2fedef790e0" containerName="kube-state-metrics" containerID="cri-o://b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d" gracePeriod=30 Dec 04 14:15:54 crc kubenswrapper[4848]: I1204 14:15:54.665049 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:15:54 crc kubenswrapper[4848]: I1204 14:15:54.665436 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="d2af561d-8819-4175-89c0-4a65a661fdb6" containerName="mysqld-exporter" containerID="cri-o://52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10" gracePeriod=30 Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.118122 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.204575 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zz5x\" (UniqueName: \"kubernetes.io/projected/d644fdb6-4d43-4e46-9537-f2fedef790e0-kube-api-access-2zz5x\") pod \"d644fdb6-4d43-4e46-9537-f2fedef790e0\" (UID: \"d644fdb6-4d43-4e46-9537-f2fedef790e0\") " Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.210294 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d644fdb6-4d43-4e46-9537-f2fedef790e0-kube-api-access-2zz5x" (OuterVolumeSpecName: "kube-api-access-2zz5x") pod "d644fdb6-4d43-4e46-9537-f2fedef790e0" (UID: "d644fdb6-4d43-4e46-9537-f2fedef790e0"). InnerVolumeSpecName "kube-api-access-2zz5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.241525 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.307051 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-combined-ca-bundle\") pod \"d2af561d-8819-4175-89c0-4a65a661fdb6\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.307120 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-config-data\") pod \"d2af561d-8819-4175-89c0-4a65a661fdb6\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.307290 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvxwc\" (UniqueName: \"kubernetes.io/projected/d2af561d-8819-4175-89c0-4a65a661fdb6-kube-api-access-rvxwc\") pod \"d2af561d-8819-4175-89c0-4a65a661fdb6\" (UID: \"d2af561d-8819-4175-89c0-4a65a661fdb6\") " Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.310130 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zz5x\" (UniqueName: \"kubernetes.io/projected/d644fdb6-4d43-4e46-9537-f2fedef790e0-kube-api-access-2zz5x\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.311057 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2af561d-8819-4175-89c0-4a65a661fdb6-kube-api-access-rvxwc" (OuterVolumeSpecName: "kube-api-access-rvxwc") pod "d2af561d-8819-4175-89c0-4a65a661fdb6" (UID: "d2af561d-8819-4175-89c0-4a65a661fdb6"). InnerVolumeSpecName "kube-api-access-rvxwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.352589 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2af561d-8819-4175-89c0-4a65a661fdb6" (UID: "d2af561d-8819-4175-89c0-4a65a661fdb6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.396827 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-config-data" (OuterVolumeSpecName: "config-data") pod "d2af561d-8819-4175-89c0-4a65a661fdb6" (UID: "d2af561d-8819-4175-89c0-4a65a661fdb6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.412258 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.412288 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2af561d-8819-4175-89c0-4a65a661fdb6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.412300 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvxwc\" (UniqueName: \"kubernetes.io/projected/d2af561d-8819-4175-89c0-4a65a661fdb6-kube-api-access-rvxwc\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.444074 4848 generic.go:334] "Generic (PLEG): container finished" podID="d2af561d-8819-4175-89c0-4a65a661fdb6" containerID="52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10" exitCode=2 Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.444141 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.444167 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"d2af561d-8819-4175-89c0-4a65a661fdb6","Type":"ContainerDied","Data":"52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10"} Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.444195 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"d2af561d-8819-4175-89c0-4a65a661fdb6","Type":"ContainerDied","Data":"c97d3f879a997959064e5310eb04bb6c1cbf55a4328493f9ab33cec3d34900f8"} Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.444213 4848 scope.go:117] "RemoveContainer" containerID="52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.448795 4848 generic.go:334] "Generic (PLEG): container finished" podID="d644fdb6-4d43-4e46-9537-f2fedef790e0" containerID="b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d" exitCode=2 Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.448841 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d644fdb6-4d43-4e46-9537-f2fedef790e0","Type":"ContainerDied","Data":"b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d"} Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.448861 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d644fdb6-4d43-4e46-9537-f2fedef790e0","Type":"ContainerDied","Data":"f0933b6194a199f8f21e481395def7a57d5dd68b9b1c03c32ed20d2a1b50b399"} Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.448866 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.482093 4848 scope.go:117] "RemoveContainer" containerID="52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10" Dec 04 14:15:55 crc kubenswrapper[4848]: E1204 14:15:55.489249 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10\": container with ID starting with 52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10 not found: ID does not exist" containerID="52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.489298 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10"} err="failed to get container status \"52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10\": rpc error: code = NotFound desc = could not find container \"52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10\": container with ID starting with 52aadde9ba602cfc441220af9f8e85894f179bbb35b6727916d9f1672d667e10 not found: ID does not exist" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.489327 4848 scope.go:117] "RemoveContainer" containerID="b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.503589 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.516251 4848 scope.go:117] "RemoveContainer" containerID="b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d" Dec 04 14:15:55 crc kubenswrapper[4848]: E1204 14:15:55.516862 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d\": container with ID starting with b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d not found: ID does not exist" containerID="b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.517011 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d"} err="failed to get container status \"b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d\": rpc error: code = NotFound desc = could not find container \"b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d\": container with ID starting with b59f604653835393755b671a5a2d1e901b96f631a4df5497016e281ef243772d not found: ID does not exist" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.519913 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.541743 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.566029 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.576215 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:15:55 crc kubenswrapper[4848]: E1204 14:15:55.576820 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d644fdb6-4d43-4e46-9537-f2fedef790e0" containerName="kube-state-metrics" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.576847 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d644fdb6-4d43-4e46-9537-f2fedef790e0" containerName="kube-state-metrics" Dec 04 14:15:55 crc kubenswrapper[4848]: E1204 14:15:55.576872 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2af561d-8819-4175-89c0-4a65a661fdb6" containerName="mysqld-exporter" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.576882 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2af561d-8819-4175-89c0-4a65a661fdb6" containerName="mysqld-exporter" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.577216 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d644fdb6-4d43-4e46-9537-f2fedef790e0" containerName="kube-state-metrics" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.577244 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2af561d-8819-4175-89c0-4a65a661fdb6" containerName="mysqld-exporter" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.578384 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.583493 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.584134 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.584525 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.594909 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.596912 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.604354 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.604562 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.606899 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.724271 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.724387 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.724804 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.724916 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pmf9\" (UniqueName: \"kubernetes.io/projected/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-kube-api-access-6pmf9\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.724965 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhll4\" (UniqueName: \"kubernetes.io/projected/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-api-access-lhll4\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.725008 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.725059 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-config-data\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.725105 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.828035 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.828462 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.828590 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.828642 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pmf9\" (UniqueName: \"kubernetes.io/projected/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-kube-api-access-6pmf9\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.828663 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhll4\" (UniqueName: \"kubernetes.io/projected/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-api-access-lhll4\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.828713 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.828760 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-config-data\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.828845 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.834996 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.836223 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.839294 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-config-data\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.840561 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.843713 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.843762 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.858009 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pmf9\" (UniqueName: \"kubernetes.io/projected/2ddeb253-877f-4e3e-8d1c-b51c1621bd15-kube-api-access-6pmf9\") pod \"mysqld-exporter-0\" (UID: \"2ddeb253-877f-4e3e-8d1c-b51c1621bd15\") " pod="openstack/mysqld-exporter-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.864471 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhll4\" (UniqueName: \"kubernetes.io/projected/1d88c8b4-d31b-45d7-8ae5-99b50d716ae0-kube-api-access-lhll4\") pod \"kube-state-metrics-0\" (UID: \"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0\") " pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.924663 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 14:15:55 crc kubenswrapper[4848]: I1204 14:15:55.938159 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 04 14:15:56 crc kubenswrapper[4848]: I1204 14:15:56.417192 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2af561d-8819-4175-89c0-4a65a661fdb6" path="/var/lib/kubelet/pods/d2af561d-8819-4175-89c0-4a65a661fdb6/volumes" Dec 04 14:15:56 crc kubenswrapper[4848]: I1204 14:15:56.418087 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d644fdb6-4d43-4e46-9537-f2fedef790e0" path="/var/lib/kubelet/pods/d644fdb6-4d43-4e46-9537-f2fedef790e0/volumes" Dec 04 14:15:56 crc kubenswrapper[4848]: I1204 14:15:56.479785 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:15:56 crc kubenswrapper[4848]: I1204 14:15:56.482416 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:15:56 crc kubenswrapper[4848]: I1204 14:15:56.486827 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:15:56 crc kubenswrapper[4848]: I1204 14:15:56.651047 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:15:56 crc kubenswrapper[4848]: I1204 14:15:56.654296 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 14:15:56 crc kubenswrapper[4848]: I1204 14:15:56.808323 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.145426 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.145937 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="proxy-httpd" containerID="cri-o://52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b" gracePeriod=30 Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.145991 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="ceilometer-notification-agent" containerID="cri-o://2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91" gracePeriod=30 Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.145920 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="ceilometer-central-agent" containerID="cri-o://c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33" gracePeriod=30 Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.145981 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="sg-core" containerID="cri-o://11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a" gracePeriod=30 Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.494144 4848 generic.go:334] "Generic (PLEG): container finished" podID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerID="52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b" exitCode=0 Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.494471 4848 generic.go:334] "Generic (PLEG): container finished" podID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerID="11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a" exitCode=2 Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.494489 4848 generic.go:334] "Generic (PLEG): container finished" podID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerID="c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33" exitCode=0 Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.494221 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerDied","Data":"52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b"} Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.494604 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerDied","Data":"11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a"} Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.494629 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerDied","Data":"c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33"} Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.496074 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0","Type":"ContainerStarted","Data":"322e64ad881212314089ffb4f8ab6d09df2fa170fc02aefa3d578b8003723912"} Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.497302 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"2ddeb253-877f-4e3e-8d1c-b51c1621bd15","Type":"ContainerStarted","Data":"487c3d82362cf7e29ea57e87012c4c5feccc9847a8b7cd8fb5da06ebdff06f58"} Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.518059 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.603839 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.604508 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.615095 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:15:57 crc kubenswrapper[4848]: I1204 14:15:57.638150 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:15:58 crc kubenswrapper[4848]: I1204 14:15:58.511307 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1d88c8b4-d31b-45d7-8ae5-99b50d716ae0","Type":"ContainerStarted","Data":"b4636c81ff10e33bbc39aeb742adba66bdb998cbc09a64e73562ddd363ec26ce"} Dec 04 14:15:58 crc kubenswrapper[4848]: I1204 14:15:58.511805 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 14:15:58 crc kubenswrapper[4848]: I1204 14:15:58.513755 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"2ddeb253-877f-4e3e-8d1c-b51c1621bd15","Type":"ContainerStarted","Data":"12e4e417d8431cad1a3812b6cb0c460ac53a1999259821b0578368d3037f62c1"} Dec 04 14:15:58 crc kubenswrapper[4848]: I1204 14:15:58.513937 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:15:58 crc kubenswrapper[4848]: I1204 14:15:58.520405 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:15:58 crc kubenswrapper[4848]: I1204 14:15:58.542475 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.908367145 podStartE2EDuration="3.542451467s" podCreationTimestamp="2025-12-04 14:15:55 +0000 UTC" firstStartedPulling="2025-12-04 14:15:56.650823 +0000 UTC m=+1660.593319528" lastFinishedPulling="2025-12-04 14:15:57.284907322 +0000 UTC m=+1661.227403850" observedRunningTime="2025-12-04 14:15:58.530821424 +0000 UTC m=+1662.473317962" watchObservedRunningTime="2025-12-04 14:15:58.542451467 +0000 UTC m=+1662.484947995" Dec 04 14:15:58 crc kubenswrapper[4848]: I1204 14:15:58.563990 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.9566023169999998 podStartE2EDuration="3.563965869s" podCreationTimestamp="2025-12-04 14:15:55 +0000 UTC" firstStartedPulling="2025-12-04 14:15:56.802067151 +0000 UTC m=+1660.744563679" lastFinishedPulling="2025-12-04 14:15:57.409430703 +0000 UTC m=+1661.351927231" observedRunningTime="2025-12-04 14:15:58.55165219 +0000 UTC m=+1662.494148718" watchObservedRunningTime="2025-12-04 14:15:58.563965869 +0000 UTC m=+1662.506462407" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.330175 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.426723 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kbxf\" (UniqueName: \"kubernetes.io/projected/5c8022e0-1093-4739-84b9-c8f77b8c1dea-kube-api-access-7kbxf\") pod \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.426882 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-scripts\") pod \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.426913 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-log-httpd\") pod \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.427074 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-sg-core-conf-yaml\") pod \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.427182 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-config-data\") pod \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.427248 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-run-httpd\") pod \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.427278 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-combined-ca-bundle\") pod \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\" (UID: \"5c8022e0-1093-4739-84b9-c8f77b8c1dea\") " Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.428969 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5c8022e0-1093-4739-84b9-c8f77b8c1dea" (UID: "5c8022e0-1093-4739-84b9-c8f77b8c1dea"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.432200 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5c8022e0-1093-4739-84b9-c8f77b8c1dea" (UID: "5c8022e0-1093-4739-84b9-c8f77b8c1dea"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.450381 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c8022e0-1093-4739-84b9-c8f77b8c1dea-kube-api-access-7kbxf" (OuterVolumeSpecName: "kube-api-access-7kbxf") pod "5c8022e0-1093-4739-84b9-c8f77b8c1dea" (UID: "5c8022e0-1093-4739-84b9-c8f77b8c1dea"). InnerVolumeSpecName "kube-api-access-7kbxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.452099 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-scripts" (OuterVolumeSpecName: "scripts") pod "5c8022e0-1093-4739-84b9-c8f77b8c1dea" (UID: "5c8022e0-1093-4739-84b9-c8f77b8c1dea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.473867 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5c8022e0-1093-4739-84b9-c8f77b8c1dea" (UID: "5c8022e0-1093-4739-84b9-c8f77b8c1dea"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.527759 4848 generic.go:334] "Generic (PLEG): container finished" podID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerID="2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91" exitCode=0 Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.527854 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.527912 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerDied","Data":"2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91"} Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.527979 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5c8022e0-1093-4739-84b9-c8f77b8c1dea","Type":"ContainerDied","Data":"6ad36aeacbb5b045bb63f344c0fdb0caaa1806c3966b01fc82ea994cf09afa33"} Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.527996 4848 scope.go:117] "RemoveContainer" containerID="52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.531217 4848 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.531235 4848 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.531244 4848 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5c8022e0-1093-4739-84b9-c8f77b8c1dea-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.531253 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kbxf\" (UniqueName: \"kubernetes.io/projected/5c8022e0-1093-4739-84b9-c8f77b8c1dea-kube-api-access-7kbxf\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.531263 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.564985 4848 scope.go:117] "RemoveContainer" containerID="11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.581450 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c8022e0-1093-4739-84b9-c8f77b8c1dea" (UID: "5c8022e0-1093-4739-84b9-c8f77b8c1dea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.591398 4848 scope.go:117] "RemoveContainer" containerID="2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.626429 4848 scope.go:117] "RemoveContainer" containerID="c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.637133 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.642416 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-config-data" (OuterVolumeSpecName: "config-data") pod "5c8022e0-1093-4739-84b9-c8f77b8c1dea" (UID: "5c8022e0-1093-4739-84b9-c8f77b8c1dea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.652525 4848 scope.go:117] "RemoveContainer" containerID="52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b" Dec 04 14:15:59 crc kubenswrapper[4848]: E1204 14:15:59.652893 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b\": container with ID starting with 52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b not found: ID does not exist" containerID="52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.652938 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b"} err="failed to get container status \"52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b\": rpc error: code = NotFound desc = could not find container \"52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b\": container with ID starting with 52fe365b41545a3381ffeea2f50391bbbf0e698eaeeea0fdd9f846e481af3c5b not found: ID does not exist" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.652976 4848 scope.go:117] "RemoveContainer" containerID="11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a" Dec 04 14:15:59 crc kubenswrapper[4848]: E1204 14:15:59.653243 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a\": container with ID starting with 11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a not found: ID does not exist" containerID="11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.653269 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a"} err="failed to get container status \"11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a\": rpc error: code = NotFound desc = could not find container \"11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a\": container with ID starting with 11088c6200297ba2a6f60ef4bdf78e0d820476c58c0aff793206721b448c197a not found: ID does not exist" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.653283 4848 scope.go:117] "RemoveContainer" containerID="2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91" Dec 04 14:15:59 crc kubenswrapper[4848]: E1204 14:15:59.653546 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91\": container with ID starting with 2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91 not found: ID does not exist" containerID="2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.653567 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91"} err="failed to get container status \"2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91\": rpc error: code = NotFound desc = could not find container \"2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91\": container with ID starting with 2e94542c956b9a9f57f5bab2f0bfebdcab5cb9603a280be92aa62a98af1d6b91 not found: ID does not exist" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.653579 4848 scope.go:117] "RemoveContainer" containerID="c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33" Dec 04 14:15:59 crc kubenswrapper[4848]: E1204 14:15:59.653768 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33\": container with ID starting with c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33 not found: ID does not exist" containerID="c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.653798 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33"} err="failed to get container status \"c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33\": rpc error: code = NotFound desc = could not find container \"c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33\": container with ID starting with c37251be58284ca390708339602748b81923237d8c9284b537c0756f6cb94d33 not found: ID does not exist" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.740052 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c8022e0-1093-4739-84b9-c8f77b8c1dea-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.864820 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.876866 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.897822 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:15:59 crc kubenswrapper[4848]: E1204 14:15:59.898491 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="ceilometer-notification-agent" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.898517 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="ceilometer-notification-agent" Dec 04 14:15:59 crc kubenswrapper[4848]: E1204 14:15:59.898534 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="sg-core" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.898543 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="sg-core" Dec 04 14:15:59 crc kubenswrapper[4848]: E1204 14:15:59.898558 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="ceilometer-central-agent" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.898568 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="ceilometer-central-agent" Dec 04 14:15:59 crc kubenswrapper[4848]: E1204 14:15:59.898606 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="proxy-httpd" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.898614 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="proxy-httpd" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.899134 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="ceilometer-central-agent" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.899166 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="proxy-httpd" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.899189 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="sg-core" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.899212 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" containerName="ceilometer-notification-agent" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.902247 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.907033 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.907275 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.907528 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:15:59 crc kubenswrapper[4848]: I1204 14:15:59.922257 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.046517 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-run-httpd\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.046586 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.046752 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.046876 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-config-data\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.046995 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbrjs\" (UniqueName: \"kubernetes.io/projected/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-kube-api-access-bbrjs\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.047103 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.047143 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-log-httpd\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.047356 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-scripts\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.149760 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-config-data\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.149841 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbrjs\" (UniqueName: \"kubernetes.io/projected/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-kube-api-access-bbrjs\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.149906 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.149965 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-log-httpd\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.150020 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-scripts\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.150068 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-run-httpd\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.150120 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.150201 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.151297 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-log-httpd\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.151383 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-run-httpd\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.154755 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.155669 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-config-data\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.156407 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.158491 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-scripts\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.174664 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.178523 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbrjs\" (UniqueName: \"kubernetes.io/projected/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-kube-api-access-bbrjs\") pod \"ceilometer-0\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.284802 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.419499 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c8022e0-1093-4739-84b9-c8f77b8c1dea" path="/var/lib/kubelet/pods/5c8022e0-1093-4739-84b9-c8f77b8c1dea/volumes" Dec 04 14:16:00 crc kubenswrapper[4848]: I1204 14:16:00.827323 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:16:00 crc kubenswrapper[4848]: W1204 14:16:00.830998 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod220ad1a9_a105_4cc4_ad42_0d5c5a075d86.slice/crio-2d304e024eadedf45b911fdd33e252c9b374b8e368b896cadde996829d6cafd2 WatchSource:0}: Error finding container 2d304e024eadedf45b911fdd33e252c9b374b8e368b896cadde996829d6cafd2: Status 404 returned error can't find the container with id 2d304e024eadedf45b911fdd33e252c9b374b8e368b896cadde996829d6cafd2 Dec 04 14:16:01 crc kubenswrapper[4848]: I1204 14:16:01.562053 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerStarted","Data":"24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635"} Dec 04 14:16:01 crc kubenswrapper[4848]: I1204 14:16:01.562714 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerStarted","Data":"2d304e024eadedf45b911fdd33e252c9b374b8e368b896cadde996829d6cafd2"} Dec 04 14:16:02 crc kubenswrapper[4848]: I1204 14:16:02.573577 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerStarted","Data":"25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a"} Dec 04 14:16:03 crc kubenswrapper[4848]: I1204 14:16:03.588498 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerStarted","Data":"214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d"} Dec 04 14:16:05 crc kubenswrapper[4848]: I1204 14:16:05.610458 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerStarted","Data":"206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a"} Dec 04 14:16:05 crc kubenswrapper[4848]: I1204 14:16:05.611093 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:16:05 crc kubenswrapper[4848]: I1204 14:16:05.635854 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.9995978450000003 podStartE2EDuration="6.63583531s" podCreationTimestamp="2025-12-04 14:15:59 +0000 UTC" firstStartedPulling="2025-12-04 14:16:00.834300719 +0000 UTC m=+1664.776797247" lastFinishedPulling="2025-12-04 14:16:04.470538184 +0000 UTC m=+1668.413034712" observedRunningTime="2025-12-04 14:16:05.629416924 +0000 UTC m=+1669.571913452" watchObservedRunningTime="2025-12-04 14:16:05.63583531 +0000 UTC m=+1669.578331838" Dec 04 14:16:05 crc kubenswrapper[4848]: I1204 14:16:05.960581 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 14:16:06 crc kubenswrapper[4848]: I1204 14:16:06.423803 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:16:06 crc kubenswrapper[4848]: E1204 14:16:06.424568 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:16:21 crc kubenswrapper[4848]: I1204 14:16:21.394196 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:16:21 crc kubenswrapper[4848]: E1204 14:16:21.394966 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:16:30 crc kubenswrapper[4848]: I1204 14:16:30.302762 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 14:16:35 crc kubenswrapper[4848]: I1204 14:16:35.394084 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:16:35 crc kubenswrapper[4848]: E1204 14:16:35.394742 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:16:41 crc kubenswrapper[4848]: I1204 14:16:41.711110 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-ggd6n"] Dec 04 14:16:41 crc kubenswrapper[4848]: I1204 14:16:41.724597 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-ggd6n"] Dec 04 14:16:41 crc kubenswrapper[4848]: I1204 14:16:41.821097 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-9h9bt"] Dec 04 14:16:41 crc kubenswrapper[4848]: I1204 14:16:41.822880 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:41 crc kubenswrapper[4848]: I1204 14:16:41.830354 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-9h9bt"] Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.000330 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4zpv\" (UniqueName: \"kubernetes.io/projected/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-kube-api-access-l4zpv\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.000414 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-config-data\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.000683 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-combined-ca-bundle\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.103427 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4zpv\" (UniqueName: \"kubernetes.io/projected/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-kube-api-access-l4zpv\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.103914 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-config-data\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.104208 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-combined-ca-bundle\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.112426 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-config-data\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.118880 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-combined-ca-bundle\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.122708 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4zpv\" (UniqueName: \"kubernetes.io/projected/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-kube-api-access-l4zpv\") pod \"heat-db-sync-9h9bt\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.197874 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9h9bt" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.418278 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dba8b0d2-85d2-45c6-9612-0652e5dfdd33" path="/var/lib/kubelet/pods/dba8b0d2-85d2-45c6-9612-0652e5dfdd33/volumes" Dec 04 14:16:42 crc kubenswrapper[4848]: I1204 14:16:42.714349 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-9h9bt"] Dec 04 14:16:43 crc kubenswrapper[4848]: I1204 14:16:43.103888 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9h9bt" event={"ID":"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa","Type":"ContainerStarted","Data":"31845a197dbf0493d08c4d67906f2db788ae2a5f85149c1012f138b130399585"} Dec 04 14:16:44 crc kubenswrapper[4848]: I1204 14:16:44.289742 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:16:44 crc kubenswrapper[4848]: I1204 14:16:44.290359 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="ceilometer-central-agent" containerID="cri-o://24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635" gracePeriod=30 Dec 04 14:16:44 crc kubenswrapper[4848]: I1204 14:16:44.290467 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="ceilometer-notification-agent" containerID="cri-o://25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a" gracePeriod=30 Dec 04 14:16:44 crc kubenswrapper[4848]: I1204 14:16:44.290481 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="proxy-httpd" containerID="cri-o://206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a" gracePeriod=30 Dec 04 14:16:44 crc kubenswrapper[4848]: I1204 14:16:44.290486 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="sg-core" containerID="cri-o://214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d" gracePeriod=30 Dec 04 14:16:44 crc kubenswrapper[4848]: I1204 14:16:44.651157 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:16:44 crc kubenswrapper[4848]: I1204 14:16:44.728292 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:16:45 crc kubenswrapper[4848]: I1204 14:16:45.136877 4848 generic.go:334] "Generic (PLEG): container finished" podID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerID="206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a" exitCode=0 Dec 04 14:16:45 crc kubenswrapper[4848]: I1204 14:16:45.136916 4848 generic.go:334] "Generic (PLEG): container finished" podID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerID="214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d" exitCode=2 Dec 04 14:16:45 crc kubenswrapper[4848]: I1204 14:16:45.136927 4848 generic.go:334] "Generic (PLEG): container finished" podID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerID="24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635" exitCode=0 Dec 04 14:16:45 crc kubenswrapper[4848]: I1204 14:16:45.136970 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerDied","Data":"206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a"} Dec 04 14:16:45 crc kubenswrapper[4848]: I1204 14:16:45.137003 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerDied","Data":"214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d"} Dec 04 14:16:45 crc kubenswrapper[4848]: I1204 14:16:45.137017 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerDied","Data":"24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635"} Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.021104 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.111726 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbrjs\" (UniqueName: \"kubernetes.io/projected/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-kube-api-access-bbrjs\") pod \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.111765 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-scripts\") pod \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.111792 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-run-httpd\") pod \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.111837 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-combined-ca-bundle\") pod \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.111865 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-ceilometer-tls-certs\") pod \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.111940 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-log-httpd\") pod \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.112065 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-sg-core-conf-yaml\") pod \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.112082 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-config-data\") pod \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\" (UID: \"220ad1a9-a105-4cc4-ad42-0d5c5a075d86\") " Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.116134 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "220ad1a9-a105-4cc4-ad42-0d5c5a075d86" (UID: "220ad1a9-a105-4cc4-ad42-0d5c5a075d86"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.118908 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "220ad1a9-a105-4cc4-ad42-0d5c5a075d86" (UID: "220ad1a9-a105-4cc4-ad42-0d5c5a075d86"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.123592 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-scripts" (OuterVolumeSpecName: "scripts") pod "220ad1a9-a105-4cc4-ad42-0d5c5a075d86" (UID: "220ad1a9-a105-4cc4-ad42-0d5c5a075d86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.152335 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-kube-api-access-bbrjs" (OuterVolumeSpecName: "kube-api-access-bbrjs") pod "220ad1a9-a105-4cc4-ad42-0d5c5a075d86" (UID: "220ad1a9-a105-4cc4-ad42-0d5c5a075d86"). InnerVolumeSpecName "kube-api-access-bbrjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.171287 4848 generic.go:334] "Generic (PLEG): container finished" podID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerID="25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a" exitCode=0 Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.171339 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerDied","Data":"25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a"} Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.171370 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"220ad1a9-a105-4cc4-ad42-0d5c5a075d86","Type":"ContainerDied","Data":"2d304e024eadedf45b911fdd33e252c9b374b8e368b896cadde996829d6cafd2"} Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.171388 4848 scope.go:117] "RemoveContainer" containerID="206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.171604 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.237196 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbrjs\" (UniqueName: \"kubernetes.io/projected/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-kube-api-access-bbrjs\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.245423 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "220ad1a9-a105-4cc4-ad42-0d5c5a075d86" (UID: "220ad1a9-a105-4cc4-ad42-0d5c5a075d86"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.252077 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.252134 4848 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.252151 4848 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.308789 4848 scope.go:117] "RemoveContainer" containerID="214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.311783 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "220ad1a9-a105-4cc4-ad42-0d5c5a075d86" (UID: "220ad1a9-a105-4cc4-ad42-0d5c5a075d86"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.314366 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "220ad1a9-a105-4cc4-ad42-0d5c5a075d86" (UID: "220ad1a9-a105-4cc4-ad42-0d5c5a075d86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.354787 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.354844 4848 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.354854 4848 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.357209 4848 scope.go:117] "RemoveContainer" containerID="25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.390082 4848 scope.go:117] "RemoveContainer" containerID="24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.404617 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.404938 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.418736 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-config-data" (OuterVolumeSpecName: "config-data") pod "220ad1a9-a105-4cc4-ad42-0d5c5a075d86" (UID: "220ad1a9-a105-4cc4-ad42-0d5c5a075d86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.426517 4848 scope.go:117] "RemoveContainer" containerID="206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a" Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.431398 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a\": container with ID starting with 206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a not found: ID does not exist" containerID="206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.431448 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a"} err="failed to get container status \"206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a\": rpc error: code = NotFound desc = could not find container \"206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a\": container with ID starting with 206997eb86dde3e6f499a43a39df8f11ddf1068831519217ad09a114588abc0a not found: ID does not exist" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.431477 4848 scope.go:117] "RemoveContainer" containerID="214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d" Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.431817 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d\": container with ID starting with 214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d not found: ID does not exist" containerID="214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.431853 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d"} err="failed to get container status \"214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d\": rpc error: code = NotFound desc = could not find container \"214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d\": container with ID starting with 214c06dd8bfe4c150329a2a9c59217d3e447e1c9445ae4f9fe9725c6c076d86d not found: ID does not exist" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.431877 4848 scope.go:117] "RemoveContainer" containerID="25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a" Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.434042 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a\": container with ID starting with 25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a not found: ID does not exist" containerID="25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.434073 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a"} err="failed to get container status \"25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a\": rpc error: code = NotFound desc = could not find container \"25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a\": container with ID starting with 25b964fc9f98f9bf83cd26d3aa329b17c162dc371d6f38f263794c871e5e282a not found: ID does not exist" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.434091 4848 scope.go:117] "RemoveContainer" containerID="24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635" Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.434604 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635\": container with ID starting with 24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635 not found: ID does not exist" containerID="24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.434629 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635"} err="failed to get container status \"24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635\": rpc error: code = NotFound desc = could not find container \"24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635\": container with ID starting with 24a90a22858578d67b6b555e7a42fca47e315c85182c42c238ac6a11dae47635 not found: ID does not exist" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.461703 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/220ad1a9-a105-4cc4-ad42-0d5c5a075d86-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.512979 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.538802 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.557751 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.558334 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="sg-core" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.558372 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="sg-core" Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.558410 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="ceilometer-central-agent" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.558419 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="ceilometer-central-agent" Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.558440 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="ceilometer-notification-agent" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.558448 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="ceilometer-notification-agent" Dec 04 14:16:46 crc kubenswrapper[4848]: E1204 14:16:46.558473 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="proxy-httpd" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.558481 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="proxy-httpd" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.558788 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="proxy-httpd" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.558817 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="ceilometer-notification-agent" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.558834 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="sg-core" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.558862 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" containerName="ceilometer-central-agent" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.562393 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.566266 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.566464 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.566576 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.612800 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.666850 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-scripts\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.666899 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-config-data\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.666980 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvx6\" (UniqueName: \"kubernetes.io/projected/40c4de59-893b-42be-b67d-f3f067b410ac-kube-api-access-qsvx6\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.667032 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.667064 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40c4de59-893b-42be-b67d-f3f067b410ac-run-httpd\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.667155 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40c4de59-893b-42be-b67d-f3f067b410ac-log-httpd\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.667195 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.667267 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.769631 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvx6\" (UniqueName: \"kubernetes.io/projected/40c4de59-893b-42be-b67d-f3f067b410ac-kube-api-access-qsvx6\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.769728 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.769781 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40c4de59-893b-42be-b67d-f3f067b410ac-run-httpd\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.769838 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40c4de59-893b-42be-b67d-f3f067b410ac-log-httpd\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.769882 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.770010 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.770203 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-scripts\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.770258 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-config-data\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.770422 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40c4de59-893b-42be-b67d-f3f067b410ac-log-httpd\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.770555 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40c4de59-893b-42be-b67d-f3f067b410ac-run-httpd\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.774153 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.774413 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.775916 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-scripts\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.787882 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-config-data\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.789504 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40c4de59-893b-42be-b67d-f3f067b410ac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.790408 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvx6\" (UniqueName: \"kubernetes.io/projected/40c4de59-893b-42be-b67d-f3f067b410ac-kube-api-access-qsvx6\") pod \"ceilometer-0\" (UID: \"40c4de59-893b-42be-b67d-f3f067b410ac\") " pod="openstack/ceilometer-0" Dec 04 14:16:46 crc kubenswrapper[4848]: I1204 14:16:46.939791 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 14:16:47 crc kubenswrapper[4848]: I1204 14:16:47.568592 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 14:16:47 crc kubenswrapper[4848]: W1204 14:16:47.618859 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40c4de59_893b_42be_b67d_f3f067b410ac.slice/crio-e59f59b8334b61f2b5242f9a7918a3831e764a44f910593edbf982ff72d56dcf WatchSource:0}: Error finding container e59f59b8334b61f2b5242f9a7918a3831e764a44f910593edbf982ff72d56dcf: Status 404 returned error can't find the container with id e59f59b8334b61f2b5242f9a7918a3831e764a44f910593edbf982ff72d56dcf Dec 04 14:16:48 crc kubenswrapper[4848]: I1204 14:16:48.217417 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40c4de59-893b-42be-b67d-f3f067b410ac","Type":"ContainerStarted","Data":"e59f59b8334b61f2b5242f9a7918a3831e764a44f910593edbf982ff72d56dcf"} Dec 04 14:16:48 crc kubenswrapper[4848]: I1204 14:16:48.408757 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="220ad1a9-a105-4cc4-ad42-0d5c5a075d86" path="/var/lib/kubelet/pods/220ad1a9-a105-4cc4-ad42-0d5c5a075d86/volumes" Dec 04 14:16:50 crc kubenswrapper[4848]: I1204 14:16:50.125816 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerName="rabbitmq" containerID="cri-o://bfa96d34d9ae2fa019286f6eeeaf829b92a08bf46a37c957eb9adf61e53823d7" gracePeriod=604795 Dec 04 14:16:50 crc kubenswrapper[4848]: I1204 14:16:50.138358 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerName="rabbitmq" containerID="cri-o://11db54d0028a98c6804adef44dccc01f2bbd27f43a15550ce6f0f8f2a8af3d45" gracePeriod=604795 Dec 04 14:16:57 crc kubenswrapper[4848]: I1204 14:16:57.621201 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.128:5671: connect: connection refused" Dec 04 14:16:57 crc kubenswrapper[4848]: I1204 14:16:57.924762 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Dec 04 14:17:00 crc kubenswrapper[4848]: I1204 14:17:00.378715 4848 generic.go:334] "Generic (PLEG): container finished" podID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerID="bfa96d34d9ae2fa019286f6eeeaf829b92a08bf46a37c957eb9adf61e53823d7" exitCode=0 Dec 04 14:17:00 crc kubenswrapper[4848]: I1204 14:17:00.378857 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30883c46-53ec-45f9-84f6-29a64ccf8994","Type":"ContainerDied","Data":"bfa96d34d9ae2fa019286f6eeeaf829b92a08bf46a37c957eb9adf61e53823d7"} Dec 04 14:17:00 crc kubenswrapper[4848]: I1204 14:17:00.385863 4848 generic.go:334] "Generic (PLEG): container finished" podID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerID="11db54d0028a98c6804adef44dccc01f2bbd27f43a15550ce6f0f8f2a8af3d45" exitCode=0 Dec 04 14:17:00 crc kubenswrapper[4848]: I1204 14:17:00.385918 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6049ed43-40db-4bff-b89b-88cce285fd2d","Type":"ContainerDied","Data":"11db54d0028a98c6804adef44dccc01f2bbd27f43a15550ce6f0f8f2a8af3d45"} Dec 04 14:17:01 crc kubenswrapper[4848]: I1204 14:17:01.394878 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:17:01 crc kubenswrapper[4848]: E1204 14:17:01.395721 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:17:04 crc kubenswrapper[4848]: I1204 14:17:04.942058 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:17:04 crc kubenswrapper[4848]: I1204 14:17:04.953431 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049546 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6049ed43-40db-4bff-b89b-88cce285fd2d-erlang-cookie-secret\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049657 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-erlang-cookie\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049693 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-config-data\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049767 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-tls\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049852 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-confd\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049890 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-server-conf\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049914 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049959 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6049ed43-40db-4bff-b89b-88cce285fd2d-pod-info\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.049993 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-plugins-conf\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.050044 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zg8j\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-kube-api-access-8zg8j\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.050068 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-plugins\") pod \"6049ed43-40db-4bff-b89b-88cce285fd2d\" (UID: \"6049ed43-40db-4bff-b89b-88cce285fd2d\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.051624 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.053521 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.053732 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.061158 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6049ed43-40db-4bff-b89b-88cce285fd2d-pod-info" (OuterVolumeSpecName: "pod-info") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.063848 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.067234 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-kube-api-access-8zg8j" (OuterVolumeSpecName: "kube-api-access-8zg8j") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "kube-api-access-8zg8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.070122 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6049ed43-40db-4bff-b89b-88cce285fd2d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.071281 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.107361 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-config-data" (OuterVolumeSpecName: "config-data") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.151617 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-plugins-conf\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.151719 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-tls\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.151763 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30883c46-53ec-45f9-84f6-29a64ccf8994-erlang-cookie-secret\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.151850 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-confd\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.151887 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.151923 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-server-conf\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.152290 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-config-data\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.152391 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kss47\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-kube-api-access-kss47\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.152434 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-plugins\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.152466 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30883c46-53ec-45f9-84f6-29a64ccf8994-pod-info\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.152516 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-erlang-cookie\") pod \"30883c46-53ec-45f9-84f6-29a64ccf8994\" (UID: \"30883c46-53ec-45f9-84f6-29a64ccf8994\") " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.153170 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156441 4848 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6049ed43-40db-4bff-b89b-88cce285fd2d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156474 4848 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156597 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156607 4848 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156615 4848 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156620 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156652 4848 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156673 4848 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6049ed43-40db-4bff-b89b-88cce285fd2d-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156682 4848 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156691 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zg8j\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-kube-api-access-8zg8j\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.156701 4848 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.157834 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30883c46-53ec-45f9-84f6-29a64ccf8994-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.159812 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/30883c46-53ec-45f9-84f6-29a64ccf8994-pod-info" (OuterVolumeSpecName: "pod-info") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.161242 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.161937 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-kube-api-access-kss47" (OuterVolumeSpecName: "kube-api-access-kss47") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "kube-api-access-kss47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.163780 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.173174 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-server-conf" (OuterVolumeSpecName: "server-conf") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.182614 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.276981 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kss47\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-kube-api-access-kss47\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.277014 4848 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.277029 4848 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30883c46-53ec-45f9-84f6-29a64ccf8994-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.277047 4848 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.277056 4848 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.277066 4848 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30883c46-53ec-45f9-84f6-29a64ccf8994-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.277076 4848 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6049ed43-40db-4bff-b89b-88cce285fd2d-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.277111 4848 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.284715 4848 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.286714 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-server-conf" (OuterVolumeSpecName: "server-conf") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.305434 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6049ed43-40db-4bff-b89b-88cce285fd2d" (UID: "6049ed43-40db-4bff-b89b-88cce285fd2d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.314060 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-config-data" (OuterVolumeSpecName: "config-data") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.323932 4848 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.379605 4848 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.379912 4848 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.380032 4848 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.380111 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30883c46-53ec-45f9-84f6-29a64ccf8994-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.380201 4848 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6049ed43-40db-4bff-b89b-88cce285fd2d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.429303 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "30883c46-53ec-45f9-84f6-29a64ccf8994" (UID: "30883c46-53ec-45f9-84f6-29a64ccf8994"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.467392 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6049ed43-40db-4bff-b89b-88cce285fd2d","Type":"ContainerDied","Data":"b2e2421f5614069784bc785cfd645ef5c68ba2126d050bdd27b12f15716d4f92"} Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.467544 4848 scope.go:117] "RemoveContainer" containerID="11db54d0028a98c6804adef44dccc01f2bbd27f43a15550ce6f0f8f2a8af3d45" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.467786 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.473591 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30883c46-53ec-45f9-84f6-29a64ccf8994","Type":"ContainerDied","Data":"f0f6cd6caa4cc643f3b4098b4205579b53c6ef21e612987957414c0e257cc923"} Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.473665 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.482357 4848 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30883c46-53ec-45f9-84f6-29a64ccf8994-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.520262 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.540313 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.560266 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.578623 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.590045 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:17:05 crc kubenswrapper[4848]: E1204 14:17:05.590613 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerName="rabbitmq" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.590631 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerName="rabbitmq" Dec 04 14:17:05 crc kubenswrapper[4848]: E1204 14:17:05.590643 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerName="rabbitmq" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.590649 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerName="rabbitmq" Dec 04 14:17:05 crc kubenswrapper[4848]: E1204 14:17:05.590668 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerName="setup-container" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.590675 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerName="setup-container" Dec 04 14:17:05 crc kubenswrapper[4848]: E1204 14:17:05.590721 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerName="setup-container" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.590729 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerName="setup-container" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.590980 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="30883c46-53ec-45f9-84f6-29a64ccf8994" containerName="rabbitmq" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.590994 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="6049ed43-40db-4bff-b89b-88cce285fd2d" containerName="rabbitmq" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.592774 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.597851 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.597972 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.598025 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.598103 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.598154 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.598240 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.598577 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-lv446" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.602002 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.603981 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.613613 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.613797 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.613918 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.614266 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.614375 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.614483 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.614612 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.614734 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-xdqv6" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.631422 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.685772 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686055 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a33214fe-5596-499a-9563-0931c49821c3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686075 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686097 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686151 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686195 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686238 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686255 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686276 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686307 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686332 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686349 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686363 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686382 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz8lx\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-kube-api-access-wz8lx\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686404 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6c94605c-497c-47b2-b606-22f72e1bd0cc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686422 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a33214fe-5596-499a-9563-0931c49821c3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686455 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686470 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686507 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6c94605c-497c-47b2-b606-22f72e1bd0cc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686524 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swhpn\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-kube-api-access-swhpn\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686559 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-config-data\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.686576 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.788996 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789094 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789123 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789152 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789196 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789231 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789254 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789259 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789274 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789301 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz8lx\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-kube-api-access-wz8lx\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789334 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6c94605c-497c-47b2-b606-22f72e1bd0cc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789361 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a33214fe-5596-499a-9563-0931c49821c3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789409 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789429 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789482 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6c94605c-497c-47b2-b606-22f72e1bd0cc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789509 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swhpn\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-kube-api-access-swhpn\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789539 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-config-data\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789560 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789622 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789647 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a33214fe-5596-499a-9563-0931c49821c3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789670 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789695 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789744 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.789913 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.790624 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.790624 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.790701 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.791286 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-config-data\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.791602 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.791916 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.793007 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6c94605c-497c-47b2-b606-22f72e1bd0cc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.793290 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.794153 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a33214fe-5596-499a-9563-0931c49821c3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.795524 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6c94605c-497c-47b2-b606-22f72e1bd0cc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.796727 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a33214fe-5596-499a-9563-0931c49821c3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.797044 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.798885 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.799486 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.799630 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.802702 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a33214fe-5596-499a-9563-0931c49821c3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.803851 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6c94605c-497c-47b2-b606-22f72e1bd0cc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.804557 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.805572 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz8lx\" (UniqueName: \"kubernetes.io/projected/6c94605c-497c-47b2-b606-22f72e1bd0cc-kube-api-access-wz8lx\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.815356 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swhpn\" (UniqueName: \"kubernetes.io/projected/a33214fe-5596-499a-9563-0931c49821c3-kube-api-access-swhpn\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.833810 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6c94605c-497c-47b2-b606-22f72e1bd0cc\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.839916 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"a33214fe-5596-499a-9563-0931c49821c3\") " pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.926702 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 14:17:05 crc kubenswrapper[4848]: I1204 14:17:05.940514 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:17:06 crc kubenswrapper[4848]: I1204 14:17:06.411084 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30883c46-53ec-45f9-84f6-29a64ccf8994" path="/var/lib/kubelet/pods/30883c46-53ec-45f9-84f6-29a64ccf8994/volumes" Dec 04 14:17:06 crc kubenswrapper[4848]: I1204 14:17:06.413506 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6049ed43-40db-4bff-b89b-88cce285fd2d" path="/var/lib/kubelet/pods/6049ed43-40db-4bff-b89b-88cce285fd2d/volumes" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.792397 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-vndhh"] Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.794704 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.798489 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.845012 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-vndhh"] Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.912395 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-vndhh"] Dec 04 14:17:07 crc kubenswrapper[4848]: E1204 14:17:07.913331 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-2d4xf openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" podUID="5345ec72-8cee-4afb-af94-2a2492af0d15" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.954133 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-g7qqg"] Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.969549 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-g7qqg"] Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.970385 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.977813 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.977893 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.978035 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.978090 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.978196 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d4xf\" (UniqueName: \"kubernetes.io/projected/5345ec72-8cee-4afb-af94-2a2492af0d15-kube-api-access-2d4xf\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.978238 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:07 crc kubenswrapper[4848]: I1204 14:17:07.978433 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-config\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080633 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080685 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080733 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080762 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d4xf\" (UniqueName: \"kubernetes.io/projected/5345ec72-8cee-4afb-af94-2a2492af0d15-kube-api-access-2d4xf\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080789 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080835 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080857 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-config\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080883 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.080979 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-config\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.081033 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.081061 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.081100 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.081136 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q48vh\" (UniqueName: \"kubernetes.io/projected/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-kube-api-access-q48vh\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.081152 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.081787 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.081873 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.082426 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-config\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.082571 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.083043 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.083306 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.101693 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d4xf\" (UniqueName: \"kubernetes.io/projected/5345ec72-8cee-4afb-af94-2a2492af0d15-kube-api-access-2d4xf\") pod \"dnsmasq-dns-7d84b4d45c-vndhh\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.183527 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-config\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.183686 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.183743 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q48vh\" (UniqueName: \"kubernetes.io/projected/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-kube-api-access-q48vh\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.183788 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.183839 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.183894 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.183938 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.185175 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-config\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.185206 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.185265 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.185328 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.186449 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.186467 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.212101 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q48vh\" (UniqueName: \"kubernetes.io/projected/7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd-kube-api-access-q48vh\") pod \"dnsmasq-dns-6f6df4f56c-g7qqg\" (UID: \"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd\") " pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.296547 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.513079 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.530554 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.695190 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-sb\") pod \"5345ec72-8cee-4afb-af94-2a2492af0d15\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.695394 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-svc\") pod \"5345ec72-8cee-4afb-af94-2a2492af0d15\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.695432 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-config\") pod \"5345ec72-8cee-4afb-af94-2a2492af0d15\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.695532 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-openstack-edpm-ipam\") pod \"5345ec72-8cee-4afb-af94-2a2492af0d15\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.695641 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5345ec72-8cee-4afb-af94-2a2492af0d15" (UID: "5345ec72-8cee-4afb-af94-2a2492af0d15"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.695661 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-swift-storage-0\") pod \"5345ec72-8cee-4afb-af94-2a2492af0d15\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.695736 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4xf\" (UniqueName: \"kubernetes.io/projected/5345ec72-8cee-4afb-af94-2a2492af0d15-kube-api-access-2d4xf\") pod \"5345ec72-8cee-4afb-af94-2a2492af0d15\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.695816 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-nb\") pod \"5345ec72-8cee-4afb-af94-2a2492af0d15\" (UID: \"5345ec72-8cee-4afb-af94-2a2492af0d15\") " Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.696105 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5345ec72-8cee-4afb-af94-2a2492af0d15" (UID: "5345ec72-8cee-4afb-af94-2a2492af0d15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.696166 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5345ec72-8cee-4afb-af94-2a2492af0d15" (UID: "5345ec72-8cee-4afb-af94-2a2492af0d15"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.696390 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5345ec72-8cee-4afb-af94-2a2492af0d15" (UID: "5345ec72-8cee-4afb-af94-2a2492af0d15"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.696921 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.696939 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.696964 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.696974 4848 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.697228 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5345ec72-8cee-4afb-af94-2a2492af0d15" (UID: "5345ec72-8cee-4afb-af94-2a2492af0d15"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.698134 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-config" (OuterVolumeSpecName: "config") pod "5345ec72-8cee-4afb-af94-2a2492af0d15" (UID: "5345ec72-8cee-4afb-af94-2a2492af0d15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.700733 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5345ec72-8cee-4afb-af94-2a2492af0d15-kube-api-access-2d4xf" (OuterVolumeSpecName: "kube-api-access-2d4xf") pod "5345ec72-8cee-4afb-af94-2a2492af0d15" (UID: "5345ec72-8cee-4afb-af94-2a2492af0d15"). InnerVolumeSpecName "kube-api-access-2d4xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.800157 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.800211 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4xf\" (UniqueName: \"kubernetes.io/projected/5345ec72-8cee-4afb-af94-2a2492af0d15-kube-api-access-2d4xf\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:08 crc kubenswrapper[4848]: I1204 14:17:08.800224 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5345ec72-8cee-4afb-af94-2a2492af0d15-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:09 crc kubenswrapper[4848]: I1204 14:17:09.530112 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-vndhh" Dec 04 14:17:09 crc kubenswrapper[4848]: I1204 14:17:09.621816 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-vndhh"] Dec 04 14:17:09 crc kubenswrapper[4848]: I1204 14:17:09.635565 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-vndhh"] Dec 04 14:17:10 crc kubenswrapper[4848]: I1204 14:17:10.410428 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5345ec72-8cee-4afb-af94-2a2492af0d15" path="/var/lib/kubelet/pods/5345ec72-8cee-4afb-af94-2a2492af0d15/volumes" Dec 04 14:17:14 crc kubenswrapper[4848]: I1204 14:17:14.395479 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd46t68x" podUID="a60c0a88-4168-4971-8bdd-638030697efb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:17:15 crc kubenswrapper[4848]: I1204 14:17:15.393168 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:17:15 crc kubenswrapper[4848]: E1204 14:17:15.393595 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:17:18 crc kubenswrapper[4848]: E1204 14:17:18.368824 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 04 14:17:18 crc kubenswrapper[4848]: E1204 14:17:18.369558 4848 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 04 14:17:18 crc kubenswrapper[4848]: E1204 14:17:18.369700 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l4zpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-9h9bt_openstack(cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:17:18 crc kubenswrapper[4848]: E1204 14:17:18.371087 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-9h9bt" podUID="cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" Dec 04 14:17:18 crc kubenswrapper[4848]: E1204 14:17:18.633414 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-9h9bt" podUID="cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" Dec 04 14:17:18 crc kubenswrapper[4848]: I1204 14:17:18.710353 4848 scope.go:117] "RemoveContainer" containerID="c4829738b3394f57da24a7580486aa1ca47db4fc2c06fe141440e1c598436e14" Dec 04 14:17:18 crc kubenswrapper[4848]: E1204 14:17:18.727758 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 04 14:17:18 crc kubenswrapper[4848]: E1204 14:17:18.727818 4848 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 04 14:17:18 crc kubenswrapper[4848]: E1204 14:17:18.728013 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d9h96h5f8hbbh54hd4h5fhchb7h576h548h66h5c9hb8h5d7h697hf4h8chdfh7bh58ch554h598h5bbh58dh6bhb6hbchc5h8dh578hfdq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qsvx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(40c4de59-893b-42be-b67d-f3f067b410ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:17:18 crc kubenswrapper[4848]: I1204 14:17:18.805899 4848 scope.go:117] "RemoveContainer" containerID="bfa96d34d9ae2fa019286f6eeeaf829b92a08bf46a37c957eb9adf61e53823d7" Dec 04 14:17:18 crc kubenswrapper[4848]: I1204 14:17:18.963325 4848 scope.go:117] "RemoveContainer" containerID="0f318664aee27c7d28309de918f73ef54729735695edff04d26277043b9fe4de" Dec 04 14:17:19 crc kubenswrapper[4848]: I1204 14:17:19.308729 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 14:17:19 crc kubenswrapper[4848]: I1204 14:17:19.345072 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 14:17:19 crc kubenswrapper[4848]: I1204 14:17:19.480153 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-g7qqg"] Dec 04 14:17:19 crc kubenswrapper[4848]: I1204 14:17:19.643367 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" event={"ID":"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd","Type":"ContainerStarted","Data":"bd566078a5fea101c1c5961d5293844da4f3fe14f0d83ad4b8a372f418be3fe9"} Dec 04 14:17:19 crc kubenswrapper[4848]: I1204 14:17:19.644636 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6c94605c-497c-47b2-b606-22f72e1bd0cc","Type":"ContainerStarted","Data":"c1415f5986df41c151da97e042625deed10a67cfe306f6438cdc91eb644335b3"} Dec 04 14:17:19 crc kubenswrapper[4848]: I1204 14:17:19.645940 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a33214fe-5596-499a-9563-0931c49821c3","Type":"ContainerStarted","Data":"934644a19f7f972365b48679ee67cb0e932626c9ce3095201f7c3390fb88d52c"} Dec 04 14:17:19 crc kubenswrapper[4848]: I1204 14:17:19.647500 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40c4de59-893b-42be-b67d-f3f067b410ac","Type":"ContainerStarted","Data":"068eb4dc41c90050832c7f8bc33067b500768cbec8b0d82399de8fba7ab52a28"} Dec 04 14:17:20 crc kubenswrapper[4848]: I1204 14:17:20.662555 4848 generic.go:334] "Generic (PLEG): container finished" podID="7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd" containerID="7ed9da9e6fbb1942e0fcfd05c8eee7624987cc4a2839d756297cb4028d3a0dfc" exitCode=0 Dec 04 14:17:20 crc kubenswrapper[4848]: I1204 14:17:20.662685 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" event={"ID":"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd","Type":"ContainerDied","Data":"7ed9da9e6fbb1942e0fcfd05c8eee7624987cc4a2839d756297cb4028d3a0dfc"} Dec 04 14:17:20 crc kubenswrapper[4848]: I1204 14:17:20.669165 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40c4de59-893b-42be-b67d-f3f067b410ac","Type":"ContainerStarted","Data":"c51c46fafee71e1aaf46901926fe040e2ce03253367618b00d3f7534dd94e465"} Dec 04 14:17:26 crc kubenswrapper[4848]: I1204 14:17:26.429335 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:17:26 crc kubenswrapper[4848]: E1204 14:17:26.431301 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:17:29 crc kubenswrapper[4848]: E1204 14:17:29.344173 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="40c4de59-893b-42be-b67d-f3f067b410ac" Dec 04 14:17:29 crc kubenswrapper[4848]: I1204 14:17:29.800155 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40c4de59-893b-42be-b67d-f3f067b410ac","Type":"ContainerStarted","Data":"37f9f7d587a389526d28e806e76e914435128aedf1b5137b603cb1f496968c31"} Dec 04 14:17:29 crc kubenswrapper[4848]: I1204 14:17:29.800304 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 14:17:29 crc kubenswrapper[4848]: I1204 14:17:29.803218 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" event={"ID":"7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd","Type":"ContainerStarted","Data":"a740ab73eb003b388a918b5b99395b4f5a51bc5439b78a93b6ee1311ff2f2250"} Dec 04 14:17:29 crc kubenswrapper[4848]: I1204 14:17:29.803307 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:29 crc kubenswrapper[4848]: I1204 14:17:29.805665 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6c94605c-497c-47b2-b606-22f72e1bd0cc","Type":"ContainerStarted","Data":"72f16736df0b122f218ae07bae916391ac87c22d7fd66d63e848ec3832ec7b9a"} Dec 04 14:17:29 crc kubenswrapper[4848]: I1204 14:17:29.807917 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a33214fe-5596-499a-9563-0931c49821c3","Type":"ContainerStarted","Data":"7bb65e1e48042b4e859cf493c4a3dfb7686c32c2110d306b5001068f8852b8a8"} Dec 04 14:17:29 crc kubenswrapper[4848]: I1204 14:17:29.911665 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" podStartSLOduration=22.911646763 podStartE2EDuration="22.911646763s" podCreationTimestamp="2025-12-04 14:17:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:29.909599413 +0000 UTC m=+1753.852095961" watchObservedRunningTime="2025-12-04 14:17:29.911646763 +0000 UTC m=+1753.854143291" Dec 04 14:17:31 crc kubenswrapper[4848]: I1204 14:17:31.844369 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40c4de59-893b-42be-b67d-f3f067b410ac","Type":"ContainerStarted","Data":"33115250cfc3815a581c98e4e55431ecfb07be7a0bd783817825dd1b7591abd6"} Dec 04 14:17:31 crc kubenswrapper[4848]: I1204 14:17:31.886242 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.829314894 podStartE2EDuration="45.886211283s" podCreationTimestamp="2025-12-04 14:16:46 +0000 UTC" firstStartedPulling="2025-12-04 14:16:47.623463427 +0000 UTC m=+1711.565959955" lastFinishedPulling="2025-12-04 14:17:30.680359816 +0000 UTC m=+1754.622856344" observedRunningTime="2025-12-04 14:17:31.867880758 +0000 UTC m=+1755.810377306" watchObservedRunningTime="2025-12-04 14:17:31.886211283 +0000 UTC m=+1755.828707821" Dec 04 14:17:33 crc kubenswrapper[4848]: I1204 14:17:33.298725 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-g7qqg" Dec 04 14:17:33 crc kubenswrapper[4848]: I1204 14:17:33.402652 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x"] Dec 04 14:17:33 crc kubenswrapper[4848]: I1204 14:17:33.403504 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" podUID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" containerName="dnsmasq-dns" containerID="cri-o://5dff0ba863b6b1e82c544163b5adc8e9eeb466523f3f4d479c9f4c39e21e0c95" gracePeriod=10 Dec 04 14:17:33 crc kubenswrapper[4848]: I1204 14:17:33.707269 4848 scope.go:117] "RemoveContainer" containerID="1e568a7ab9c8b37d95568cc59c1b6e0a0a7d7328b565bd47cb8c987ae3780f0d" Dec 04 14:17:33 crc kubenswrapper[4848]: I1204 14:17:33.767251 4848 scope.go:117] "RemoveContainer" containerID="b79aef1f5f288e7a71d74b5d217ffb3f69e3654849b2a7e05b2a2c3f8491200e" Dec 04 14:17:33 crc kubenswrapper[4848]: I1204 14:17:33.844737 4848 scope.go:117] "RemoveContainer" containerID="e8e23d771c6fa0161054cefb3e13fc1e598648961438525837ca0dced9b5ddcf" Dec 04 14:17:33 crc kubenswrapper[4848]: I1204 14:17:33.918006 4848 generic.go:334] "Generic (PLEG): container finished" podID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" containerID="5dff0ba863b6b1e82c544163b5adc8e9eeb466523f3f4d479c9f4c39e21e0c95" exitCode=0 Dec 04 14:17:33 crc kubenswrapper[4848]: I1204 14:17:33.918197 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" event={"ID":"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345","Type":"ContainerDied","Data":"5dff0ba863b6b1e82c544163b5adc8e9eeb466523f3f4d479c9f4c39e21e0c95"} Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.072909 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.224985 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-nb\") pod \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.225464 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-sb\") pod \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.225877 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-svc\") pod \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.225971 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-swift-storage-0\") pod \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.225994 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fg84\" (UniqueName: \"kubernetes.io/projected/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-kube-api-access-9fg84\") pod \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.226123 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-config\") pod \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\" (UID: \"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345\") " Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.231083 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-kube-api-access-9fg84" (OuterVolumeSpecName: "kube-api-access-9fg84") pod "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" (UID: "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345"). InnerVolumeSpecName "kube-api-access-9fg84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.300068 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" (UID: "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.301877 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-config" (OuterVolumeSpecName: "config") pod "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" (UID: "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.302708 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" (UID: "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.303239 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" (UID: "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.303279 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" (UID: "8bdf08cd-02ba-46ff-a7f6-cf14bdff5345"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.330315 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.330369 4848 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.330385 4848 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.330399 4848 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.330415 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fg84\" (UniqueName: \"kubernetes.io/projected/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-kube-api-access-9fg84\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.330429 4848 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345-config\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.940591 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9h9bt" event={"ID":"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa","Type":"ContainerStarted","Data":"e8e46ba09e90621e97561c2ff1278ce17460794bbe14aef860e25f72e8596e6c"} Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.951569 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" event={"ID":"8bdf08cd-02ba-46ff-a7f6-cf14bdff5345","Type":"ContainerDied","Data":"877f95dea8816ac327f1f68bb1b39017c4b5b41c6b1dff2f9b83a495869f9eca"} Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.951628 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.951640 4848 scope.go:117] "RemoveContainer" containerID="5dff0ba863b6b1e82c544163b5adc8e9eeb466523f3f4d479c9f4c39e21e0c95" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.966939 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-9h9bt" podStartSLOduration=2.888478631 podStartE2EDuration="53.966918887s" podCreationTimestamp="2025-12-04 14:16:41 +0000 UTC" firstStartedPulling="2025-12-04 14:16:42.715149334 +0000 UTC m=+1706.657645862" lastFinishedPulling="2025-12-04 14:17:33.79358959 +0000 UTC m=+1757.736086118" observedRunningTime="2025-12-04 14:17:34.964695464 +0000 UTC m=+1758.907191992" watchObservedRunningTime="2025-12-04 14:17:34.966918887 +0000 UTC m=+1758.909415405" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.991643 4848 scope.go:117] "RemoveContainer" containerID="c4be3ebf0c729a807282a42f023525b83585e04a0484d49371a2bc772dd9e6c6" Dec 04 14:17:34 crc kubenswrapper[4848]: I1204 14:17:34.998658 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x"] Dec 04 14:17:35 crc kubenswrapper[4848]: I1204 14:17:35.013349 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-6xv4x"] Dec 04 14:17:36 crc kubenswrapper[4848]: I1204 14:17:36.411237 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" path="/var/lib/kubelet/pods/8bdf08cd-02ba-46ff-a7f6-cf14bdff5345/volumes" Dec 04 14:17:36 crc kubenswrapper[4848]: I1204 14:17:36.983539 4848 generic.go:334] "Generic (PLEG): container finished" podID="cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" containerID="e8e46ba09e90621e97561c2ff1278ce17460794bbe14aef860e25f72e8596e6c" exitCode=0 Dec 04 14:17:36 crc kubenswrapper[4848]: I1204 14:17:36.983604 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9h9bt" event={"ID":"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa","Type":"ContainerDied","Data":"e8e46ba09e90621e97561c2ff1278ce17460794bbe14aef860e25f72e8596e6c"} Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.461663 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9h9bt" Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.645088 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4zpv\" (UniqueName: \"kubernetes.io/projected/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-kube-api-access-l4zpv\") pod \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.645240 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-config-data\") pod \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.645312 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-combined-ca-bundle\") pod \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\" (UID: \"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa\") " Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.653245 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-kube-api-access-l4zpv" (OuterVolumeSpecName: "kube-api-access-l4zpv") pod "cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" (UID: "cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa"). InnerVolumeSpecName "kube-api-access-l4zpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.687818 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" (UID: "cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.731615 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-config-data" (OuterVolumeSpecName: "config-data") pod "cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" (UID: "cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.752498 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4zpv\" (UniqueName: \"kubernetes.io/projected/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-kube-api-access-l4zpv\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.752594 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:38 crc kubenswrapper[4848]: I1204 14:17:38.752611 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:17:39 crc kubenswrapper[4848]: I1204 14:17:39.032134 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9h9bt" event={"ID":"cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa","Type":"ContainerDied","Data":"31845a197dbf0493d08c4d67906f2db788ae2a5f85149c1012f138b130399585"} Dec 04 14:17:39 crc kubenswrapper[4848]: I1204 14:17:39.032171 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31845a197dbf0493d08c4d67906f2db788ae2a5f85149c1012f138b130399585" Dec 04 14:17:39 crc kubenswrapper[4848]: I1204 14:17:39.032226 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9h9bt" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.472530 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-55957f9948-8gknv"] Dec 04 14:17:40 crc kubenswrapper[4848]: E1204 14:17:40.473418 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" containerName="heat-db-sync" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.473433 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" containerName="heat-db-sync" Dec 04 14:17:40 crc kubenswrapper[4848]: E1204 14:17:40.473464 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" containerName="dnsmasq-dns" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.473470 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" containerName="dnsmasq-dns" Dec 04 14:17:40 crc kubenswrapper[4848]: E1204 14:17:40.473482 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" containerName="init" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.473488 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" containerName="init" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.473782 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bdf08cd-02ba-46ff-a7f6-cf14bdff5345" containerName="dnsmasq-dns" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.473803 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" containerName="heat-db-sync" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.474665 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.490721 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55957f9948-8gknv"] Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.538041 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6c746cbd45-fc8bq"] Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.540167 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.594268 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-config-data\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.594403 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-config-data-custom\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.594495 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxh9k\" (UniqueName: \"kubernetes.io/projected/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-kube-api-access-mxh9k\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.594541 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-combined-ca-bundle\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.620528 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6c746cbd45-fc8bq"] Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.690017 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6b8f46df8d-vvlcd"] Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.692338 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.707662 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-config-data-custom\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.707827 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jzfh\" (UniqueName: \"kubernetes.io/projected/4ab14888-fa03-4bbd-a567-95e9078fbe3b-kube-api-access-9jzfh\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.707871 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-config-data-custom\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.707894 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-internal-tls-certs\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.708001 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-combined-ca-bundle\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.708091 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxh9k\" (UniqueName: \"kubernetes.io/projected/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-kube-api-access-mxh9k\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.708161 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-combined-ca-bundle\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.708201 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-config-data\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.708253 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-config-data\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.708308 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-public-tls-certs\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.758603 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-combined-ca-bundle\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.759393 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-config-data-custom\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.769062 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6b8f46df8d-vvlcd"] Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.778744 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-config-data\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.780738 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxh9k\" (UniqueName: \"kubernetes.io/projected/01ecc04b-d6b6-4182-ba10-0b2d14b6bef5-kube-api-access-mxh9k\") pod \"heat-engine-55957f9948-8gknv\" (UID: \"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5\") " pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.846563 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-combined-ca-bundle\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.846642 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jzfh\" (UniqueName: \"kubernetes.io/projected/4ab14888-fa03-4bbd-a567-95e9078fbe3b-kube-api-access-9jzfh\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.846694 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-internal-tls-certs\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.846785 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-internal-tls-certs\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.846830 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-public-tls-certs\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.846856 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-combined-ca-bundle\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.854673 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.856785 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-config-data\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.856887 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-config-data\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.857018 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-public-tls-certs\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.857056 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5snhw\" (UniqueName: \"kubernetes.io/projected/cfffd13e-a23e-41bd-b830-0feab48494ec-kube-api-access-5snhw\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.857083 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-config-data-custom\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.857113 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-config-data-custom\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.866717 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-internal-tls-certs\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.867876 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-config-data\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.869049 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-public-tls-certs\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.875458 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-combined-ca-bundle\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.878324 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ab14888-fa03-4bbd-a567-95e9078fbe3b-config-data-custom\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.909944 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jzfh\" (UniqueName: \"kubernetes.io/projected/4ab14888-fa03-4bbd-a567-95e9078fbe3b-kube-api-access-9jzfh\") pod \"heat-api-6c746cbd45-fc8bq\" (UID: \"4ab14888-fa03-4bbd-a567-95e9078fbe3b\") " pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.960745 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-internal-tls-certs\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.960800 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-public-tls-certs\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.960900 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-config-data\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.961050 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5snhw\" (UniqueName: \"kubernetes.io/projected/cfffd13e-a23e-41bd-b830-0feab48494ec-kube-api-access-5snhw\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.961079 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-config-data-custom\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.961178 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-combined-ca-bundle\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.967783 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-internal-tls-certs\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.969870 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-config-data\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.972138 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-combined-ca-bundle\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.977282 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-public-tls-certs\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.986366 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfffd13e-a23e-41bd-b830-0feab48494ec-config-data-custom\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:40 crc kubenswrapper[4848]: I1204 14:17:40.988432 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5snhw\" (UniqueName: \"kubernetes.io/projected/cfffd13e-a23e-41bd-b830-0feab48494ec-kube-api-access-5snhw\") pod \"heat-cfnapi-6b8f46df8d-vvlcd\" (UID: \"cfffd13e-a23e-41bd-b830-0feab48494ec\") " pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:41 crc kubenswrapper[4848]: I1204 14:17:41.107235 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:41 crc kubenswrapper[4848]: I1204 14:17:41.166381 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:41 crc kubenswrapper[4848]: I1204 14:17:41.393758 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:17:41 crc kubenswrapper[4848]: E1204 14:17:41.394272 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:17:41 crc kubenswrapper[4848]: I1204 14:17:41.426232 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55957f9948-8gknv"] Dec 04 14:17:41 crc kubenswrapper[4848]: I1204 14:17:41.622485 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6b8f46df8d-vvlcd"] Dec 04 14:17:41 crc kubenswrapper[4848]: W1204 14:17:41.779993 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ab14888_fa03_4bbd_a567_95e9078fbe3b.slice/crio-fc2a19ec8674e2da34e6ebf3fc137a6e0f14adc9dfca95da7883ba02384c59db WatchSource:0}: Error finding container fc2a19ec8674e2da34e6ebf3fc137a6e0f14adc9dfca95da7883ba02384c59db: Status 404 returned error can't find the container with id fc2a19ec8674e2da34e6ebf3fc137a6e0f14adc9dfca95da7883ba02384c59db Dec 04 14:17:41 crc kubenswrapper[4848]: I1204 14:17:41.784937 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6c746cbd45-fc8bq"] Dec 04 14:17:42 crc kubenswrapper[4848]: I1204 14:17:42.354638 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6c746cbd45-fc8bq" event={"ID":"4ab14888-fa03-4bbd-a567-95e9078fbe3b","Type":"ContainerStarted","Data":"fc2a19ec8674e2da34e6ebf3fc137a6e0f14adc9dfca95da7883ba02384c59db"} Dec 04 14:17:42 crc kubenswrapper[4848]: I1204 14:17:42.356472 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" event={"ID":"cfffd13e-a23e-41bd-b830-0feab48494ec","Type":"ContainerStarted","Data":"2fa117b03eb72f870776b06a3f1e966e61f3aed9a07378cb332485d528e06a94"} Dec 04 14:17:42 crc kubenswrapper[4848]: I1204 14:17:42.357749 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55957f9948-8gknv" event={"ID":"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5","Type":"ContainerStarted","Data":"982a01ee5864fb2b6d2673297ad7494db2fe93cc49236845351978db3561d576"} Dec 04 14:17:45 crc kubenswrapper[4848]: I1204 14:17:45.128227 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" podUID="12302e73-00be-4ffb-9cda-3ac6126e2f0c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:17:45 crc kubenswrapper[4848]: I1204 14:17:45.128608 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-576c48d787-gx8wm" podUID="12302e73-00be-4ffb-9cda-3ac6126e2f0c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:17:46 crc kubenswrapper[4848]: I1204 14:17:46.398174 4848 patch_prober.go:28] interesting pod/monitoring-plugin-97c9cddfd-672td container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.75:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 14:17:46 crc kubenswrapper[4848]: I1204 14:17:46.398472 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-97c9cddfd-672td" podUID="74119917-08ea-4ef2-801d-232f7be20625" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.75:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 14:17:46 crc kubenswrapper[4848]: I1204 14:17:46.957491 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.102942 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz"] Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.104874 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.107629 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.107694 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.107819 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.108010 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.117490 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz"] Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.231892 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vwqw\" (UniqueName: \"kubernetes.io/projected/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-kube-api-access-2vwqw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.231995 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.232040 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.232319 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.335311 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vwqw\" (UniqueName: \"kubernetes.io/projected/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-kube-api-access-2vwqw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.335416 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.336425 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.336545 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.342562 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.343202 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.351913 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.353028 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vwqw\" (UniqueName: \"kubernetes.io/projected/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-kube-api-access-2vwqw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:47 crc kubenswrapper[4848]: I1204 14:17:47.435602 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:17:48 crc kubenswrapper[4848]: I1204 14:17:48.425090 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55957f9948-8gknv" event={"ID":"01ecc04b-d6b6-4182-ba10-0b2d14b6bef5","Type":"ContainerStarted","Data":"920e553a7b9dd4189bf9fc2f7ad5906078cd15e37aa993ea75b13e26ab3a7e5d"} Dec 04 14:17:49 crc kubenswrapper[4848]: I1204 14:17:49.435912 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:17:49 crc kubenswrapper[4848]: I1204 14:17:49.452803 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-55957f9948-8gknv" podStartSLOduration=9.452787418 podStartE2EDuration="9.452787418s" podCreationTimestamp="2025-12-04 14:17:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:17:49.450416331 +0000 UTC m=+1773.392912859" watchObservedRunningTime="2025-12-04 14:17:49.452787418 +0000 UTC m=+1773.395283946" Dec 04 14:17:55 crc kubenswrapper[4848]: I1204 14:17:55.393256 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:17:55 crc kubenswrapper[4848]: E1204 14:17:55.394083 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:17:57 crc kubenswrapper[4848]: W1204 14:17:57.454400 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66ec76cb_b6b1_4846_83d3_58cf60fd2d3d.slice/crio-ab61be9f2164b9ec88db6b0154feacfb66e16f3c73e8c15ae2aadedad92863b6 WatchSource:0}: Error finding container ab61be9f2164b9ec88db6b0154feacfb66e16f3c73e8c15ae2aadedad92863b6: Status 404 returned error can't find the container with id ab61be9f2164b9ec88db6b0154feacfb66e16f3c73e8c15ae2aadedad92863b6 Dec 04 14:17:57 crc kubenswrapper[4848]: I1204 14:17:57.460191 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz"] Dec 04 14:17:57 crc kubenswrapper[4848]: I1204 14:17:57.523143 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" event={"ID":"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d","Type":"ContainerStarted","Data":"ab61be9f2164b9ec88db6b0154feacfb66e16f3c73e8c15ae2aadedad92863b6"} Dec 04 14:17:57 crc kubenswrapper[4848]: I1204 14:17:57.525237 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" event={"ID":"cfffd13e-a23e-41bd-b830-0feab48494ec","Type":"ContainerStarted","Data":"f0cc0b842f9cedcc70a2b08f2b5b5ab06e6b9d3f26ddc382495d36e0c85c5b90"} Dec 04 14:17:57 crc kubenswrapper[4848]: I1204 14:17:57.525318 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:17:57 crc kubenswrapper[4848]: I1204 14:17:57.526684 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6c746cbd45-fc8bq" event={"ID":"4ab14888-fa03-4bbd-a567-95e9078fbe3b","Type":"ContainerStarted","Data":"9c3f12581a672b2fe13aa84fbbc700481cdae68851f94ca324817c670620df6f"} Dec 04 14:17:57 crc kubenswrapper[4848]: I1204 14:17:57.526870 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:17:57 crc kubenswrapper[4848]: I1204 14:17:57.560337 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" podStartSLOduration=3.03051865 podStartE2EDuration="17.560317858s" podCreationTimestamp="2025-12-04 14:17:40 +0000 UTC" firstStartedPulling="2025-12-04 14:17:41.625367998 +0000 UTC m=+1765.567864526" lastFinishedPulling="2025-12-04 14:17:56.155167186 +0000 UTC m=+1780.097663734" observedRunningTime="2025-12-04 14:17:57.544055772 +0000 UTC m=+1781.486552300" watchObservedRunningTime="2025-12-04 14:17:57.560317858 +0000 UTC m=+1781.502814386" Dec 04 14:17:57 crc kubenswrapper[4848]: I1204 14:17:57.589705 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6c746cbd45-fc8bq" podStartSLOduration=3.217114025 podStartE2EDuration="17.589688722s" podCreationTimestamp="2025-12-04 14:17:40 +0000 UTC" firstStartedPulling="2025-12-04 14:17:41.782340583 +0000 UTC m=+1765.724837101" lastFinishedPulling="2025-12-04 14:17:56.15491527 +0000 UTC m=+1780.097411798" observedRunningTime="2025-12-04 14:17:57.569537882 +0000 UTC m=+1781.512034410" watchObservedRunningTime="2025-12-04 14:17:57.589688722 +0000 UTC m=+1781.532185250" Dec 04 14:18:00 crc kubenswrapper[4848]: I1204 14:18:00.930256 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-55957f9948-8gknv" Dec 04 14:18:01 crc kubenswrapper[4848]: I1204 14:18:01.026563 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-55dd7d4b8f-9bj88"] Dec 04 14:18:01 crc kubenswrapper[4848]: I1204 14:18:01.026806 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-55dd7d4b8f-9bj88" podUID="732df5f3-a0ba-449b-bf97-8604a569314f" containerName="heat-engine" containerID="cri-o://ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" gracePeriod=60 Dec 04 14:18:01 crc kubenswrapper[4848]: I1204 14:18:01.589375 4848 generic.go:334] "Generic (PLEG): container finished" podID="a33214fe-5596-499a-9563-0931c49821c3" containerID="7bb65e1e48042b4e859cf493c4a3dfb7686c32c2110d306b5001068f8852b8a8" exitCode=0 Dec 04 14:18:01 crc kubenswrapper[4848]: I1204 14:18:01.589634 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a33214fe-5596-499a-9563-0931c49821c3","Type":"ContainerDied","Data":"7bb65e1e48042b4e859cf493c4a3dfb7686c32c2110d306b5001068f8852b8a8"} Dec 04 14:18:01 crc kubenswrapper[4848]: I1204 14:18:01.610111 4848 generic.go:334] "Generic (PLEG): container finished" podID="6c94605c-497c-47b2-b606-22f72e1bd0cc" containerID="72f16736df0b122f218ae07bae916391ac87c22d7fd66d63e848ec3832ec7b9a" exitCode=0 Dec 04 14:18:01 crc kubenswrapper[4848]: I1204 14:18:01.610154 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6c94605c-497c-47b2-b606-22f72e1bd0cc","Type":"ContainerDied","Data":"72f16736df0b122f218ae07bae916391ac87c22d7fd66d63e848ec3832ec7b9a"} Dec 04 14:18:02 crc kubenswrapper[4848]: I1204 14:18:02.622133 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a33214fe-5596-499a-9563-0931c49821c3","Type":"ContainerStarted","Data":"f90e1401c8b818d0c8caad050a52572b9f5a6583f3454acf4ea3770ffacb6eb4"} Dec 04 14:18:02 crc kubenswrapper[4848]: I1204 14:18:02.622806 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 14:18:02 crc kubenswrapper[4848]: I1204 14:18:02.625251 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6c94605c-497c-47b2-b606-22f72e1bd0cc","Type":"ContainerStarted","Data":"95798e45bdce3689559f7da91a6003138bd545f86a02cab521b0fce2d6da6a71"} Dec 04 14:18:02 crc kubenswrapper[4848]: I1204 14:18:02.625800 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:18:02 crc kubenswrapper[4848]: I1204 14:18:02.668082 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=57.668060089 podStartE2EDuration="57.668060089s" podCreationTimestamp="2025-12-04 14:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:18:02.64919751 +0000 UTC m=+1786.591694038" watchObservedRunningTime="2025-12-04 14:18:02.668060089 +0000 UTC m=+1786.610556617" Dec 04 14:18:02 crc kubenswrapper[4848]: I1204 14:18:02.702813 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=57.702791313 podStartE2EDuration="57.702791313s" podCreationTimestamp="2025-12-04 14:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:18:02.685322338 +0000 UTC m=+1786.627818866" watchObservedRunningTime="2025-12-04 14:18:02.702791313 +0000 UTC m=+1786.645287841" Dec 04 14:18:04 crc kubenswrapper[4848]: I1204 14:18:04.488980 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-6b8f46df8d-vvlcd" Dec 04 14:18:04 crc kubenswrapper[4848]: I1204 14:18:04.606491 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6c746cbd45-fc8bq" Dec 04 14:18:04 crc kubenswrapper[4848]: I1204 14:18:04.613538 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7fcddc797c-lm59t"] Dec 04 14:18:04 crc kubenswrapper[4848]: I1204 14:18:04.614107 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" podUID="e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" containerName="heat-cfnapi" containerID="cri-o://a74545ca7904da03f0d3b9228bbc0239cbf44b77771af84bf1a6800ed27f0255" gracePeriod=60 Dec 04 14:18:04 crc kubenswrapper[4848]: I1204 14:18:04.731783 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-79cb47cd96-sdhxb"] Dec 04 14:18:04 crc kubenswrapper[4848]: I1204 14:18:04.741527 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-79cb47cd96-sdhxb" podUID="a1b263e8-986b-4822-9d9d-f1454757cadb" containerName="heat-api" containerID="cri-o://1ee1caa64261f0dbf0c1ccd8ba30c50baaf85f0078f6881ee025818185371175" gracePeriod=60 Dec 04 14:18:06 crc kubenswrapper[4848]: E1204 14:18:06.577446 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:18:06 crc kubenswrapper[4848]: E1204 14:18:06.578821 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:18:06 crc kubenswrapper[4848]: E1204 14:18:06.579701 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:18:06 crc kubenswrapper[4848]: E1204 14:18:06.579729 4848 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-55dd7d4b8f-9bj88" podUID="732df5f3-a0ba-449b-bf97-8604a569314f" containerName="heat-engine" Dec 04 14:18:08 crc kubenswrapper[4848]: I1204 14:18:08.491001 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-79cb47cd96-sdhxb" podUID="a1b263e8-986b-4822-9d9d-f1454757cadb" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.215:8004/healthcheck\": dial tcp 10.217.0.215:8004: connect: connection refused" Dec 04 14:18:08 crc kubenswrapper[4848]: I1204 14:18:08.513582 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" podUID="e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.216:8000/healthcheck\": dial tcp 10.217.0.216:8000: connect: connection refused" Dec 04 14:18:08 crc kubenswrapper[4848]: I1204 14:18:08.737432 4848 generic.go:334] "Generic (PLEG): container finished" podID="e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" containerID="a74545ca7904da03f0d3b9228bbc0239cbf44b77771af84bf1a6800ed27f0255" exitCode=0 Dec 04 14:18:08 crc kubenswrapper[4848]: I1204 14:18:08.737510 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" event={"ID":"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9","Type":"ContainerDied","Data":"a74545ca7904da03f0d3b9228bbc0239cbf44b77771af84bf1a6800ed27f0255"} Dec 04 14:18:08 crc kubenswrapper[4848]: I1204 14:18:08.748131 4848 generic.go:334] "Generic (PLEG): container finished" podID="a1b263e8-986b-4822-9d9d-f1454757cadb" containerID="1ee1caa64261f0dbf0c1ccd8ba30c50baaf85f0078f6881ee025818185371175" exitCode=0 Dec 04 14:18:08 crc kubenswrapper[4848]: I1204 14:18:08.748188 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-79cb47cd96-sdhxb" event={"ID":"a1b263e8-986b-4822-9d9d-f1454757cadb","Type":"ContainerDied","Data":"1ee1caa64261f0dbf0c1ccd8ba30c50baaf85f0078f6881ee025818185371175"} Dec 04 14:18:09 crc kubenswrapper[4848]: I1204 14:18:09.393023 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:18:09 crc kubenswrapper[4848]: E1204 14:18:09.393510 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:18:11 crc kubenswrapper[4848]: I1204 14:18:11.933210 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-kxmlp"] Dec 04 14:18:11 crc kubenswrapper[4848]: I1204 14:18:11.945937 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-kxmlp"] Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.032773 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-6jlbh"] Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.034860 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.036738 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.046517 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6jlbh"] Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.159645 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-scripts\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.159699 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-combined-ca-bundle\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.159730 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp7fn\" (UniqueName: \"kubernetes.io/projected/cf1c92f4-2158-4669-9816-cfda5923549a-kube-api-access-bp7fn\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.159890 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-config-data\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.262021 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-config-data\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.263107 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-scripts\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.263183 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-combined-ca-bundle\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.263263 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp7fn\" (UniqueName: \"kubernetes.io/projected/cf1c92f4-2158-4669-9816-cfda5923549a-kube-api-access-bp7fn\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.286963 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-config-data\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.287291 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-scripts\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.287971 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-combined-ca-bundle\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.297678 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp7fn\" (UniqueName: \"kubernetes.io/projected/cf1c92f4-2158-4669-9816-cfda5923549a-kube-api-access-bp7fn\") pod \"aodh-db-sync-6jlbh\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.365064 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:12 crc kubenswrapper[4848]: I1204 14:18:12.409105 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="397d21bc-b1cc-49a6-b9c8-353c772cf769" path="/var/lib/kubelet/pods/397d21bc-b1cc-49a6-b9c8-353c772cf769/volumes" Dec 04 14:18:12 crc kubenswrapper[4848]: E1204 14:18:12.658877 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Dec 04 14:18:12 crc kubenswrapper[4848]: E1204 14:18:12.659073 4848 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 14:18:12 crc kubenswrapper[4848]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Dec 04 14:18:12 crc kubenswrapper[4848]: - hosts: all Dec 04 14:18:12 crc kubenswrapper[4848]: strategy: linear Dec 04 14:18:12 crc kubenswrapper[4848]: tasks: Dec 04 14:18:12 crc kubenswrapper[4848]: - name: Enable podified-repos Dec 04 14:18:12 crc kubenswrapper[4848]: become: true Dec 04 14:18:12 crc kubenswrapper[4848]: ansible.builtin.shell: | Dec 04 14:18:12 crc kubenswrapper[4848]: set -euxo pipefail Dec 04 14:18:12 crc kubenswrapper[4848]: pushd /var/tmp Dec 04 14:18:12 crc kubenswrapper[4848]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Dec 04 14:18:12 crc kubenswrapper[4848]: pushd repo-setup-main Dec 04 14:18:12 crc kubenswrapper[4848]: python3 -m venv ./venv Dec 04 14:18:12 crc kubenswrapper[4848]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Dec 04 14:18:12 crc kubenswrapper[4848]: ./venv/bin/repo-setup current-podified -b antelope Dec 04 14:18:12 crc kubenswrapper[4848]: popd Dec 04 14:18:12 crc kubenswrapper[4848]: rm -rf repo-setup-main Dec 04 14:18:12 crc kubenswrapper[4848]: Dec 04 14:18:12 crc kubenswrapper[4848]: Dec 04 14:18:12 crc kubenswrapper[4848]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Dec 04 14:18:12 crc kubenswrapper[4848]: edpm_override_hosts: openstack-edpm-ipam Dec 04 14:18:12 crc kubenswrapper[4848]: edpm_service_type: repo-setup Dec 04 14:18:12 crc kubenswrapper[4848]: Dec 04 14:18:12 crc kubenswrapper[4848]: Dec 04 14:18:12 crc kubenswrapper[4848]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2vwqw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz_openstack(66ec76cb-b6b1-4846-83d3-58cf60fd2d3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Dec 04 14:18:12 crc kubenswrapper[4848]: > logger="UnhandledError" Dec 04 14:18:12 crc kubenswrapper[4848]: E1204 14:18:12.661347 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" podUID="66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" Dec 04 14:18:12 crc kubenswrapper[4848]: E1204 14:18:12.824766 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" podUID="66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.429799 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.437692 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.492847 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6jlbh"] Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.600921 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl6wc\" (UniqueName: \"kubernetes.io/projected/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-kube-api-access-gl6wc\") pod \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601020 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r6dh\" (UniqueName: \"kubernetes.io/projected/a1b263e8-986b-4822-9d9d-f1454757cadb-kube-api-access-9r6dh\") pod \"a1b263e8-986b-4822-9d9d-f1454757cadb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601062 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data\") pod \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601127 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-public-tls-certs\") pod \"a1b263e8-986b-4822-9d9d-f1454757cadb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601165 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-combined-ca-bundle\") pod \"a1b263e8-986b-4822-9d9d-f1454757cadb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601233 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data-custom\") pod \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601261 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-combined-ca-bundle\") pod \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601364 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data-custom\") pod \"a1b263e8-986b-4822-9d9d-f1454757cadb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601506 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-public-tls-certs\") pod \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601553 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-internal-tls-certs\") pod \"a1b263e8-986b-4822-9d9d-f1454757cadb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601611 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-internal-tls-certs\") pod \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\" (UID: \"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.601654 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data\") pod \"a1b263e8-986b-4822-9d9d-f1454757cadb\" (UID: \"a1b263e8-986b-4822-9d9d-f1454757cadb\") " Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.608250 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" (UID: "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.610171 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-kube-api-access-gl6wc" (OuterVolumeSpecName: "kube-api-access-gl6wc") pod "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" (UID: "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9"). InnerVolumeSpecName "kube-api-access-gl6wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.616562 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a1b263e8-986b-4822-9d9d-f1454757cadb" (UID: "a1b263e8-986b-4822-9d9d-f1454757cadb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.620133 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b263e8-986b-4822-9d9d-f1454757cadb-kube-api-access-9r6dh" (OuterVolumeSpecName: "kube-api-access-9r6dh") pod "a1b263e8-986b-4822-9d9d-f1454757cadb" (UID: "a1b263e8-986b-4822-9d9d-f1454757cadb"). InnerVolumeSpecName "kube-api-access-9r6dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.658325 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" (UID: "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.666283 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1b263e8-986b-4822-9d9d-f1454757cadb" (UID: "a1b263e8-986b-4822-9d9d-f1454757cadb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.707070 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.707098 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.707108 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.707118 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.707128 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl6wc\" (UniqueName: \"kubernetes.io/projected/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-kube-api-access-gl6wc\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.707137 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r6dh\" (UniqueName: \"kubernetes.io/projected/a1b263e8-986b-4822-9d9d-f1454757cadb-kube-api-access-9r6dh\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.712504 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a1b263e8-986b-4822-9d9d-f1454757cadb" (UID: "a1b263e8-986b-4822-9d9d-f1454757cadb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.712590 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" (UID: "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.720182 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data" (OuterVolumeSpecName: "config-data") pod "a1b263e8-986b-4822-9d9d-f1454757cadb" (UID: "a1b263e8-986b-4822-9d9d-f1454757cadb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.720714 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a1b263e8-986b-4822-9d9d-f1454757cadb" (UID: "a1b263e8-986b-4822-9d9d-f1454757cadb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.725890 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data" (OuterVolumeSpecName: "config-data") pod "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" (UID: "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.743405 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" (UID: "e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.809347 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.809396 4848 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.809410 4848 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.809421 4848 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.809432 4848 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.809442 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b263e8-986b-4822-9d9d-f1454757cadb-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.832928 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6jlbh" event={"ID":"cf1c92f4-2158-4669-9816-cfda5923549a","Type":"ContainerStarted","Data":"e5ff561158ef21e854233c5670acc439f36789fdc9bc89cd531e1aeb1527b52a"} Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.834936 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-79cb47cd96-sdhxb" event={"ID":"a1b263e8-986b-4822-9d9d-f1454757cadb","Type":"ContainerDied","Data":"876713e322f8762625d5bf3b53d1f7702866228ec46495334ae74632a3dc3feb"} Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.834999 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-79cb47cd96-sdhxb" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.835031 4848 scope.go:117] "RemoveContainer" containerID="1ee1caa64261f0dbf0c1ccd8ba30c50baaf85f0078f6881ee025818185371175" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.838913 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" event={"ID":"e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9","Type":"ContainerDied","Data":"b0b68ef983defd475a7df4577f43b9ef21339fafa35e48d88b54116bd9824fe3"} Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.839018 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7fcddc797c-lm59t" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.866505 4848 scope.go:117] "RemoveContainer" containerID="a74545ca7904da03f0d3b9228bbc0239cbf44b77771af84bf1a6800ed27f0255" Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.881848 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-79cb47cd96-sdhxb"] Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.900897 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-79cb47cd96-sdhxb"] Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.916190 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7fcddc797c-lm59t"] Dec 04 14:18:13 crc kubenswrapper[4848]: I1204 14:18:13.928023 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7fcddc797c-lm59t"] Dec 04 14:18:13 crc kubenswrapper[4848]: E1204 14:18:13.952604 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1b263e8_986b_4822_9d9d_f1454757cadb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1b263e8_986b_4822_9d9d_f1454757cadb.slice/crio-876713e322f8762625d5bf3b53d1f7702866228ec46495334ae74632a3dc3feb\": RecentStats: unable to find data in memory cache]" Dec 04 14:18:14 crc kubenswrapper[4848]: I1204 14:18:14.415997 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1b263e8-986b-4822-9d9d-f1454757cadb" path="/var/lib/kubelet/pods/a1b263e8-986b-4822-9d9d-f1454757cadb/volumes" Dec 04 14:18:14 crc kubenswrapper[4848]: I1204 14:18:14.418095 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" path="/var/lib/kubelet/pods/e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9/volumes" Dec 04 14:18:15 crc kubenswrapper[4848]: I1204 14:18:15.940428 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="a33214fe-5596-499a-9563-0931c49821c3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Dec 04 14:18:15 crc kubenswrapper[4848]: I1204 14:18:15.943182 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 14:18:16 crc kubenswrapper[4848]: E1204 14:18:16.580085 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:18:16 crc kubenswrapper[4848]: E1204 14:18:16.587655 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:18:16 crc kubenswrapper[4848]: E1204 14:18:16.589232 4848 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 04 14:18:16 crc kubenswrapper[4848]: E1204 14:18:16.589301 4848 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-55dd7d4b8f-9bj88" podUID="732df5f3-a0ba-449b-bf97-8604a569314f" containerName="heat-engine" Dec 04 14:18:19 crc kubenswrapper[4848]: I1204 14:18:19.356688 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 14:18:19 crc kubenswrapper[4848]: I1204 14:18:19.938593 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6jlbh" event={"ID":"cf1c92f4-2158-4669-9816-cfda5923549a","Type":"ContainerStarted","Data":"967008c19d4c8753fdcb441e40778498f8829fd91e80e8babe4b842c6d36816e"} Dec 04 14:18:19 crc kubenswrapper[4848]: I1204 14:18:19.965112 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-6jlbh" podStartSLOduration=2.176426583 podStartE2EDuration="7.965093154s" podCreationTimestamp="2025-12-04 14:18:12 +0000 UTC" firstStartedPulling="2025-12-04 14:18:13.564331616 +0000 UTC m=+1797.506828144" lastFinishedPulling="2025-12-04 14:18:19.352998187 +0000 UTC m=+1803.295494715" observedRunningTime="2025-12-04 14:18:19.957209092 +0000 UTC m=+1803.899705620" watchObservedRunningTime="2025-12-04 14:18:19.965093154 +0000 UTC m=+1803.907589682" Dec 04 14:18:20 crc kubenswrapper[4848]: I1204 14:18:20.393883 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:18:20 crc kubenswrapper[4848]: E1204 14:18:20.394773 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:18:22 crc kubenswrapper[4848]: I1204 14:18:22.981635 4848 generic.go:334] "Generic (PLEG): container finished" podID="732df5f3-a0ba-449b-bf97-8604a569314f" containerID="ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" exitCode=0 Dec 04 14:18:22 crc kubenswrapper[4848]: I1204 14:18:22.981793 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55dd7d4b8f-9bj88" event={"ID":"732df5f3-a0ba-449b-bf97-8604a569314f","Type":"ContainerDied","Data":"ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1"} Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.827866 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.991150 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-combined-ca-bundle\") pod \"732df5f3-a0ba-449b-bf97-8604a569314f\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.991714 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scrbj\" (UniqueName: \"kubernetes.io/projected/732df5f3-a0ba-449b-bf97-8604a569314f-kube-api-access-scrbj\") pod \"732df5f3-a0ba-449b-bf97-8604a569314f\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.991933 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data\") pod \"732df5f3-a0ba-449b-bf97-8604a569314f\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.992057 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data-custom\") pod \"732df5f3-a0ba-449b-bf97-8604a569314f\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.995526 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55dd7d4b8f-9bj88" event={"ID":"732df5f3-a0ba-449b-bf97-8604a569314f","Type":"ContainerDied","Data":"c0ec08d301c28b573439ae3ab2329665c0a216ff3bcdf8f54dbb2221b219cb08"} Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.995568 4848 scope.go:117] "RemoveContainer" containerID="ec48eed24f451bb2d2e0a519313eac2048a2a8f277a21d7381a91c124c9c27e1" Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.995682 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55dd7d4b8f-9bj88" Dec 04 14:18:23 crc kubenswrapper[4848]: I1204 14:18:23.999868 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/732df5f3-a0ba-449b-bf97-8604a569314f-kube-api-access-scrbj" (OuterVolumeSpecName: "kube-api-access-scrbj") pod "732df5f3-a0ba-449b-bf97-8604a569314f" (UID: "732df5f3-a0ba-449b-bf97-8604a569314f"). InnerVolumeSpecName "kube-api-access-scrbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.001047 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "732df5f3-a0ba-449b-bf97-8604a569314f" (UID: "732df5f3-a0ba-449b-bf97-8604a569314f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:24 crc kubenswrapper[4848]: E1204 14:18:24.055579 4848 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data podName:732df5f3-a0ba-449b-bf97-8604a569314f nodeName:}" failed. No retries permitted until 2025-12-04 14:18:24.555545219 +0000 UTC m=+1808.498041757 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data") pod "732df5f3-a0ba-449b-bf97-8604a569314f" (UID: "732df5f3-a0ba-449b-bf97-8604a569314f") : error deleting /var/lib/kubelet/pods/732df5f3-a0ba-449b-bf97-8604a569314f/volume-subpaths: remove /var/lib/kubelet/pods/732df5f3-a0ba-449b-bf97-8604a569314f/volume-subpaths: no such file or directory Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.061103 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "732df5f3-a0ba-449b-bf97-8604a569314f" (UID: "732df5f3-a0ba-449b-bf97-8604a569314f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.096164 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scrbj\" (UniqueName: \"kubernetes.io/projected/732df5f3-a0ba-449b-bf97-8604a569314f-kube-api-access-scrbj\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.096200 4848 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.096215 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.608460 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data\") pod \"732df5f3-a0ba-449b-bf97-8604a569314f\" (UID: \"732df5f3-a0ba-449b-bf97-8604a569314f\") " Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.620028 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data" (OuterVolumeSpecName: "config-data") pod "732df5f3-a0ba-449b-bf97-8604a569314f" (UID: "732df5f3-a0ba-449b-bf97-8604a569314f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.712792 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732df5f3-a0ba-449b-bf97-8604a569314f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.940192 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-55dd7d4b8f-9bj88"] Dec 04 14:18:24 crc kubenswrapper[4848]: I1204 14:18:24.953402 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-55dd7d4b8f-9bj88"] Dec 04 14:18:25 crc kubenswrapper[4848]: I1204 14:18:25.011530 4848 generic.go:334] "Generic (PLEG): container finished" podID="cf1c92f4-2158-4669-9816-cfda5923549a" containerID="967008c19d4c8753fdcb441e40778498f8829fd91e80e8babe4b842c6d36816e" exitCode=0 Dec 04 14:18:25 crc kubenswrapper[4848]: I1204 14:18:25.011567 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6jlbh" event={"ID":"cf1c92f4-2158-4669-9816-cfda5923549a","Type":"ContainerDied","Data":"967008c19d4c8753fdcb441e40778498f8829fd91e80e8babe4b842c6d36816e"} Dec 04 14:18:25 crc kubenswrapper[4848]: I1204 14:18:25.930443 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.416401 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="732df5f3-a0ba-449b-bf97-8604a569314f" path="/var/lib/kubelet/pods/732df5f3-a0ba-449b-bf97-8604a569314f/volumes" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.514739 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.602314 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp7fn\" (UniqueName: \"kubernetes.io/projected/cf1c92f4-2158-4669-9816-cfda5923549a-kube-api-access-bp7fn\") pod \"cf1c92f4-2158-4669-9816-cfda5923549a\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.602531 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-combined-ca-bundle\") pod \"cf1c92f4-2158-4669-9816-cfda5923549a\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.602686 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-scripts\") pod \"cf1c92f4-2158-4669-9816-cfda5923549a\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.602734 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-config-data\") pod \"cf1c92f4-2158-4669-9816-cfda5923549a\" (UID: \"cf1c92f4-2158-4669-9816-cfda5923549a\") " Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.609715 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-scripts" (OuterVolumeSpecName: "scripts") pod "cf1c92f4-2158-4669-9816-cfda5923549a" (UID: "cf1c92f4-2158-4669-9816-cfda5923549a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.618272 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf1c92f4-2158-4669-9816-cfda5923549a-kube-api-access-bp7fn" (OuterVolumeSpecName: "kube-api-access-bp7fn") pod "cf1c92f4-2158-4669-9816-cfda5923549a" (UID: "cf1c92f4-2158-4669-9816-cfda5923549a"). InnerVolumeSpecName "kube-api-access-bp7fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.642784 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf1c92f4-2158-4669-9816-cfda5923549a" (UID: "cf1c92f4-2158-4669-9816-cfda5923549a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.644393 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-config-data" (OuterVolumeSpecName: "config-data") pod "cf1c92f4-2158-4669-9816-cfda5923549a" (UID: "cf1c92f4-2158-4669-9816-cfda5923549a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.706465 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp7fn\" (UniqueName: \"kubernetes.io/projected/cf1c92f4-2158-4669-9816-cfda5923549a-kube-api-access-bp7fn\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.706759 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.706821 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:26 crc kubenswrapper[4848]: I1204 14:18:26.706888 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1c92f4-2158-4669-9816-cfda5923549a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:27 crc kubenswrapper[4848]: I1204 14:18:27.043839 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6jlbh" event={"ID":"cf1c92f4-2158-4669-9816-cfda5923549a","Type":"ContainerDied","Data":"e5ff561158ef21e854233c5670acc439f36789fdc9bc89cd531e1aeb1527b52a"} Dec 04 14:18:27 crc kubenswrapper[4848]: I1204 14:18:27.044114 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5ff561158ef21e854233c5670acc439f36789fdc9bc89cd531e1aeb1527b52a" Dec 04 14:18:27 crc kubenswrapper[4848]: I1204 14:18:27.043913 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6jlbh" Dec 04 14:18:27 crc kubenswrapper[4848]: I1204 14:18:27.370000 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:18:28 crc kubenswrapper[4848]: I1204 14:18:28.057791 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" event={"ID":"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d","Type":"ContainerStarted","Data":"f0940f8ae938e69b2674baa72003f3e07145422a8fad9b650008e68a8b8fdae2"} Dec 04 14:18:28 crc kubenswrapper[4848]: I1204 14:18:28.073123 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" podStartSLOduration=11.163013647 podStartE2EDuration="41.073105325s" podCreationTimestamp="2025-12-04 14:17:47 +0000 UTC" firstStartedPulling="2025-12-04 14:17:57.456965846 +0000 UTC m=+1781.399462374" lastFinishedPulling="2025-12-04 14:18:27.367057524 +0000 UTC m=+1811.309554052" observedRunningTime="2025-12-04 14:18:28.071172937 +0000 UTC m=+1812.013669485" watchObservedRunningTime="2025-12-04 14:18:28.073105325 +0000 UTC m=+1812.015601853" Dec 04 14:18:31 crc kubenswrapper[4848]: I1204 14:18:31.393707 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:18:31 crc kubenswrapper[4848]: E1204 14:18:31.394737 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:18:31 crc kubenswrapper[4848]: I1204 14:18:31.997542 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 04 14:18:31 crc kubenswrapper[4848]: I1204 14:18:31.997813 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-api" containerID="cri-o://8d7254c20441ec1adb4671df1a7577740de25dea51bc7a57bfab8e723a317206" gracePeriod=30 Dec 04 14:18:31 crc kubenswrapper[4848]: I1204 14:18:31.997922 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-evaluator" containerID="cri-o://7ef085a9fae995c05f374d9e2f2c080fc23a75bd76296891538d2733e97fbe2d" gracePeriod=30 Dec 04 14:18:31 crc kubenswrapper[4848]: I1204 14:18:31.997910 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-notifier" containerID="cri-o://f70f58c1a5d10fa2785184800d1b2560c76a9b5c4f17647a20f542ef2845c975" gracePeriod=30 Dec 04 14:18:31 crc kubenswrapper[4848]: I1204 14:18:31.998029 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-listener" containerID="cri-o://16e37ee5614acb0246bb4311be23a7b27c15fbba434517217cda5ab58abec44f" gracePeriod=30 Dec 04 14:18:33 crc kubenswrapper[4848]: I1204 14:18:33.119709 4848 generic.go:334] "Generic (PLEG): container finished" podID="a570d885-c380-4d05-af91-890fa9eb893b" containerID="7ef085a9fae995c05f374d9e2f2c080fc23a75bd76296891538d2733e97fbe2d" exitCode=0 Dec 04 14:18:33 crc kubenswrapper[4848]: I1204 14:18:33.120705 4848 generic.go:334] "Generic (PLEG): container finished" podID="a570d885-c380-4d05-af91-890fa9eb893b" containerID="8d7254c20441ec1adb4671df1a7577740de25dea51bc7a57bfab8e723a317206" exitCode=0 Dec 04 14:18:33 crc kubenswrapper[4848]: I1204 14:18:33.119798 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerDied","Data":"7ef085a9fae995c05f374d9e2f2c080fc23a75bd76296891538d2733e97fbe2d"} Dec 04 14:18:33 crc kubenswrapper[4848]: I1204 14:18:33.120906 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerDied","Data":"8d7254c20441ec1adb4671df1a7577740de25dea51bc7a57bfab8e723a317206"} Dec 04 14:18:34 crc kubenswrapper[4848]: I1204 14:18:34.268767 4848 scope.go:117] "RemoveContainer" containerID="34efa360e1a91ddb73a5293b1621957a6d5eaba3f5b9101dedb9b126e90776d5" Dec 04 14:18:34 crc kubenswrapper[4848]: I1204 14:18:34.314038 4848 scope.go:117] "RemoveContainer" containerID="68899fbbe0a11f00377c19314d4e5f511d7f1555f358fbe9cf6d5159de5965f1" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.155657 4848 generic.go:334] "Generic (PLEG): container finished" podID="a570d885-c380-4d05-af91-890fa9eb893b" containerID="16e37ee5614acb0246bb4311be23a7b27c15fbba434517217cda5ab58abec44f" exitCode=0 Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.155931 4848 generic.go:334] "Generic (PLEG): container finished" podID="a570d885-c380-4d05-af91-890fa9eb893b" containerID="f70f58c1a5d10fa2785184800d1b2560c76a9b5c4f17647a20f542ef2845c975" exitCode=0 Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.155739 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerDied","Data":"16e37ee5614acb0246bb4311be23a7b27c15fbba434517217cda5ab58abec44f"} Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.155998 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerDied","Data":"f70f58c1a5d10fa2785184800d1b2560c76a9b5c4f17647a20f542ef2845c975"} Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.637542 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.684067 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-internal-tls-certs\") pod \"a570d885-c380-4d05-af91-890fa9eb893b\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.684153 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-public-tls-certs\") pod \"a570d885-c380-4d05-af91-890fa9eb893b\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.684175 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-config-data\") pod \"a570d885-c380-4d05-af91-890fa9eb893b\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.684241 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-scripts\") pod \"a570d885-c380-4d05-af91-890fa9eb893b\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.684375 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpbb6\" (UniqueName: \"kubernetes.io/projected/a570d885-c380-4d05-af91-890fa9eb893b-kube-api-access-dpbb6\") pod \"a570d885-c380-4d05-af91-890fa9eb893b\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.684395 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-combined-ca-bundle\") pod \"a570d885-c380-4d05-af91-890fa9eb893b\" (UID: \"a570d885-c380-4d05-af91-890fa9eb893b\") " Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.690016 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-scripts" (OuterVolumeSpecName: "scripts") pod "a570d885-c380-4d05-af91-890fa9eb893b" (UID: "a570d885-c380-4d05-af91-890fa9eb893b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.692175 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a570d885-c380-4d05-af91-890fa9eb893b-kube-api-access-dpbb6" (OuterVolumeSpecName: "kube-api-access-dpbb6") pod "a570d885-c380-4d05-af91-890fa9eb893b" (UID: "a570d885-c380-4d05-af91-890fa9eb893b"). InnerVolumeSpecName "kube-api-access-dpbb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.807033 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpbb6\" (UniqueName: \"kubernetes.io/projected/a570d885-c380-4d05-af91-890fa9eb893b-kube-api-access-dpbb6\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.807093 4848 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.833467 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a570d885-c380-4d05-af91-890fa9eb893b" (UID: "a570d885-c380-4d05-af91-890fa9eb893b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.833523 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a570d885-c380-4d05-af91-890fa9eb893b" (UID: "a570d885-c380-4d05-af91-890fa9eb893b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.898280 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a570d885-c380-4d05-af91-890fa9eb893b" (UID: "a570d885-c380-4d05-af91-890fa9eb893b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.920711 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.920743 4848 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.920752 4848 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:35 crc kubenswrapper[4848]: I1204 14:18:35.997116 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-config-data" (OuterVolumeSpecName: "config-data") pod "a570d885-c380-4d05-af91-890fa9eb893b" (UID: "a570d885-c380-4d05-af91-890fa9eb893b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.022368 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a570d885-c380-4d05-af91-890fa9eb893b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.183268 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"a570d885-c380-4d05-af91-890fa9eb893b","Type":"ContainerDied","Data":"e5ac7a15469ee04ec9aabe605eaaaf71be05f2b012f86e0ffe3051065bad14ac"} Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.183611 4848 scope.go:117] "RemoveContainer" containerID="16e37ee5614acb0246bb4311be23a7b27c15fbba434517217cda5ab58abec44f" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.183853 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.241972 4848 scope.go:117] "RemoveContainer" containerID="f70f58c1a5d10fa2785184800d1b2560c76a9b5c4f17647a20f542ef2845c975" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.254081 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.271115 4848 scope.go:117] "RemoveContainer" containerID="7ef085a9fae995c05f374d9e2f2c080fc23a75bd76296891538d2733e97fbe2d" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.277496 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.299656 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 04 14:18:36 crc kubenswrapper[4848]: E1204 14:18:36.300330 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" containerName="heat-cfnapi" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.300417 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" containerName="heat-cfnapi" Dec 04 14:18:36 crc kubenswrapper[4848]: E1204 14:18:36.300516 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-evaluator" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.300593 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-evaluator" Dec 04 14:18:36 crc kubenswrapper[4848]: E1204 14:18:36.300675 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b263e8-986b-4822-9d9d-f1454757cadb" containerName="heat-api" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.300754 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b263e8-986b-4822-9d9d-f1454757cadb" containerName="heat-api" Dec 04 14:18:36 crc kubenswrapper[4848]: E1204 14:18:36.300826 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-api" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.300897 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-api" Dec 04 14:18:36 crc kubenswrapper[4848]: E1204 14:18:36.300986 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-listener" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.301065 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-listener" Dec 04 14:18:36 crc kubenswrapper[4848]: E1204 14:18:36.301193 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732df5f3-a0ba-449b-bf97-8604a569314f" containerName="heat-engine" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.301296 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="732df5f3-a0ba-449b-bf97-8604a569314f" containerName="heat-engine" Dec 04 14:18:36 crc kubenswrapper[4848]: E1204 14:18:36.301409 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-notifier" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.301517 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-notifier" Dec 04 14:18:36 crc kubenswrapper[4848]: E1204 14:18:36.301589 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf1c92f4-2158-4669-9816-cfda5923549a" containerName="aodh-db-sync" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.301651 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf1c92f4-2158-4669-9816-cfda5923549a" containerName="aodh-db-sync" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.301994 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-api" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.302101 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="732df5f3-a0ba-449b-bf97-8604a569314f" containerName="heat-engine" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.302376 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b263e8-986b-4822-9d9d-f1454757cadb" containerName="heat-api" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.302452 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf1c92f4-2158-4669-9816-cfda5923549a" containerName="aodh-db-sync" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.302555 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-listener" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.302644 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f9b7c8-ff8a-45ef-81e8-7dfa8e2fcbf9" containerName="heat-cfnapi" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.302722 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-evaluator" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.302788 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a570d885-c380-4d05-af91-890fa9eb893b" containerName="aodh-notifier" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.306052 4848 scope.go:117] "RemoveContainer" containerID="8d7254c20441ec1adb4671df1a7577740de25dea51bc7a57bfab8e723a317206" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.315733 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.315862 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.319558 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.319855 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.319971 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-d7qhp" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.320384 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.320960 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.408349 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a570d885-c380-4d05-af91-890fa9eb893b" path="/var/lib/kubelet/pods/a570d885-c380-4d05-af91-890fa9eb893b/volumes" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.430020 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.430362 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-scripts\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.430404 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-config-data\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.430459 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-774m2\" (UniqueName: \"kubernetes.io/projected/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-kube-api-access-774m2\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.430513 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-internal-tls-certs\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.430777 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-public-tls-certs\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.532496 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.533497 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-scripts\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.533535 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-config-data\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.533581 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-774m2\" (UniqueName: \"kubernetes.io/projected/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-kube-api-access-774m2\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.533622 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-internal-tls-certs\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.533793 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-public-tls-certs\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.537721 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-scripts\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.538478 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-internal-tls-certs\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.539079 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-public-tls-certs\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.540065 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-config-data\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.545390 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.553093 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-774m2\" (UniqueName: \"kubernetes.io/projected/bc646c01-afe4-4b6b-ab39-d8c333cc3ed4-kube-api-access-774m2\") pod \"aodh-0\" (UID: \"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4\") " pod="openstack/aodh-0" Dec 04 14:18:36 crc kubenswrapper[4848]: I1204 14:18:36.639229 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 14:18:37 crc kubenswrapper[4848]: I1204 14:18:37.126241 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 14:18:37 crc kubenswrapper[4848]: I1204 14:18:37.196843 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4","Type":"ContainerStarted","Data":"d1294f5bfcd4aace933aef1e6c3275c992d0175c90ed17806d9e9dedfd2b931f"} Dec 04 14:18:38 crc kubenswrapper[4848]: I1204 14:18:38.210427 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4","Type":"ContainerStarted","Data":"3c80e3e5ddc31f6e933ce5bc56a69f630fd6be870c4f7540d9bfb83750584056"} Dec 04 14:18:43 crc kubenswrapper[4848]: I1204 14:18:43.281574 4848 generic.go:334] "Generic (PLEG): container finished" podID="66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" containerID="f0940f8ae938e69b2674baa72003f3e07145422a8fad9b650008e68a8b8fdae2" exitCode=0 Dec 04 14:18:43 crc kubenswrapper[4848]: I1204 14:18:43.281648 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" event={"ID":"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d","Type":"ContainerDied","Data":"f0940f8ae938e69b2674baa72003f3e07145422a8fad9b650008e68a8b8fdae2"} Dec 04 14:18:44 crc kubenswrapper[4848]: I1204 14:18:44.295881 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4","Type":"ContainerStarted","Data":"424db3338511eb5aa4ae743afac5c5bd181c094f3fcbc7049759a5d3c96a5ed7"} Dec 04 14:18:44 crc kubenswrapper[4848]: I1204 14:18:44.846380 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.015505 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vwqw\" (UniqueName: \"kubernetes.io/projected/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-kube-api-access-2vwqw\") pod \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.015727 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-inventory\") pod \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.015810 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-ssh-key\") pod \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.015915 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-repo-setup-combined-ca-bundle\") pod \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\" (UID: \"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d\") " Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.021331 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-kube-api-access-2vwqw" (OuterVolumeSpecName: "kube-api-access-2vwqw") pod "66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" (UID: "66ec76cb-b6b1-4846-83d3-58cf60fd2d3d"). InnerVolumeSpecName "kube-api-access-2vwqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.023730 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" (UID: "66ec76cb-b6b1-4846-83d3-58cf60fd2d3d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.047499 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-inventory" (OuterVolumeSpecName: "inventory") pod "66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" (UID: "66ec76cb-b6b1-4846-83d3-58cf60fd2d3d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.061749 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" (UID: "66ec76cb-b6b1-4846-83d3-58cf60fd2d3d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.118681 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.118714 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.118725 4848 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.118734 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vwqw\" (UniqueName: \"kubernetes.io/projected/66ec76cb-b6b1-4846-83d3-58cf60fd2d3d-kube-api-access-2vwqw\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.319180 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" event={"ID":"66ec76cb-b6b1-4846-83d3-58cf60fd2d3d","Type":"ContainerDied","Data":"ab61be9f2164b9ec88db6b0154feacfb66e16f3c73e8c15ae2aadedad92863b6"} Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.319228 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab61be9f2164b9ec88db6b0154feacfb66e16f3c73e8c15ae2aadedad92863b6" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.319295 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.392082 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw"] Dec 04 14:18:45 crc kubenswrapper[4848]: E1204 14:18:45.392665 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.392684 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.392903 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="66ec76cb-b6b1-4846-83d3-58cf60fd2d3d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.393826 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.394546 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:18:45 crc kubenswrapper[4848]: E1204 14:18:45.395475 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.398748 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.399078 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.399200 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.399538 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.405821 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw"] Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.440894 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf8l2\" (UniqueName: \"kubernetes.io/projected/25e07c34-72be-4c3a-8432-3fb295a46d57-kube-api-access-bf8l2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.440984 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.441183 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.547386 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf8l2\" (UniqueName: \"kubernetes.io/projected/25e07c34-72be-4c3a-8432-3fb295a46d57-kube-api-access-bf8l2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.547717 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.547804 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.552840 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.553219 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.565425 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf8l2\" (UniqueName: \"kubernetes.io/projected/25e07c34-72be-4c3a-8432-3fb295a46d57-kube-api-access-bf8l2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kwkbw\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:45 crc kubenswrapper[4848]: I1204 14:18:45.717228 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:46 crc kubenswrapper[4848]: W1204 14:18:46.304171 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25e07c34_72be_4c3a_8432_3fb295a46d57.slice/crio-e706ece4b7e51d3d771b491e8fffbd4845ef5391bda1213c3861ebcde7ae57a4 WatchSource:0}: Error finding container e706ece4b7e51d3d771b491e8fffbd4845ef5391bda1213c3861ebcde7ae57a4: Status 404 returned error can't find the container with id e706ece4b7e51d3d771b491e8fffbd4845ef5391bda1213c3861ebcde7ae57a4 Dec 04 14:18:46 crc kubenswrapper[4848]: I1204 14:18:46.313554 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw"] Dec 04 14:18:46 crc kubenswrapper[4848]: I1204 14:18:46.332933 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4","Type":"ContainerStarted","Data":"8e6b32180f72feab28e7c2b617da65a01b8950d19f2a0a4037ca88419307a701"} Dec 04 14:18:46 crc kubenswrapper[4848]: I1204 14:18:46.336339 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" event={"ID":"25e07c34-72be-4c3a-8432-3fb295a46d57","Type":"ContainerStarted","Data":"e706ece4b7e51d3d771b491e8fffbd4845ef5391bda1213c3861ebcde7ae57a4"} Dec 04 14:18:48 crc kubenswrapper[4848]: I1204 14:18:48.446127 4848 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7ddb79586-x6vmg container/manager namespace/openshift-operators-redhat: Liveness probe status=failure output="Get \"http://10.217.0.49:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 14:18:48 crc kubenswrapper[4848]: I1204 14:18:48.446446 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" podUID="461d1ece-26ff-40ed-a92c-f21300adeec7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.49:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:18:50 crc kubenswrapper[4848]: I1204 14:18:50.416326 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"bc646c01-afe4-4b6b-ab39-d8c333cc3ed4","Type":"ContainerStarted","Data":"c9a196375f2c73c99f36316244763e05fa5112dfd36895a19c9e4b74da9c4e0b"} Dec 04 14:18:50 crc kubenswrapper[4848]: I1204 14:18:50.417738 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" event={"ID":"25e07c34-72be-4c3a-8432-3fb295a46d57","Type":"ContainerStarted","Data":"ce11ee8be0efea2a71baee2e18e57b25791169608d73b3b066448a41f66e384d"} Dec 04 14:18:50 crc kubenswrapper[4848]: I1204 14:18:50.453147 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.407394474 podStartE2EDuration="14.453125499s" podCreationTimestamp="2025-12-04 14:18:36 +0000 UTC" firstStartedPulling="2025-12-04 14:18:37.123561591 +0000 UTC m=+1821.066058119" lastFinishedPulling="2025-12-04 14:18:49.169292616 +0000 UTC m=+1833.111789144" observedRunningTime="2025-12-04 14:18:50.441558157 +0000 UTC m=+1834.384054685" watchObservedRunningTime="2025-12-04 14:18:50.453125499 +0000 UTC m=+1834.395622027" Dec 04 14:18:50 crc kubenswrapper[4848]: I1204 14:18:50.476906 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" podStartSLOduration=2.61702311 podStartE2EDuration="5.476877066s" podCreationTimestamp="2025-12-04 14:18:45 +0000 UTC" firstStartedPulling="2025-12-04 14:18:46.311659893 +0000 UTC m=+1830.254156461" lastFinishedPulling="2025-12-04 14:18:49.171513889 +0000 UTC m=+1833.114010417" observedRunningTime="2025-12-04 14:18:50.468456471 +0000 UTC m=+1834.410953049" watchObservedRunningTime="2025-12-04 14:18:50.476877066 +0000 UTC m=+1834.419373634" Dec 04 14:18:52 crc kubenswrapper[4848]: I1204 14:18:52.436929 4848 generic.go:334] "Generic (PLEG): container finished" podID="25e07c34-72be-4c3a-8432-3fb295a46d57" containerID="ce11ee8be0efea2a71baee2e18e57b25791169608d73b3b066448a41f66e384d" exitCode=0 Dec 04 14:18:52 crc kubenswrapper[4848]: I1204 14:18:52.437469 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" event={"ID":"25e07c34-72be-4c3a-8432-3fb295a46d57","Type":"ContainerDied","Data":"ce11ee8be0efea2a71baee2e18e57b25791169608d73b3b066448a41f66e384d"} Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.018937 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.070997 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf8l2\" (UniqueName: \"kubernetes.io/projected/25e07c34-72be-4c3a-8432-3fb295a46d57-kube-api-access-bf8l2\") pod \"25e07c34-72be-4c3a-8432-3fb295a46d57\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.071237 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-ssh-key\") pod \"25e07c34-72be-4c3a-8432-3fb295a46d57\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.071430 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-inventory\") pod \"25e07c34-72be-4c3a-8432-3fb295a46d57\" (UID: \"25e07c34-72be-4c3a-8432-3fb295a46d57\") " Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.077066 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e07c34-72be-4c3a-8432-3fb295a46d57-kube-api-access-bf8l2" (OuterVolumeSpecName: "kube-api-access-bf8l2") pod "25e07c34-72be-4c3a-8432-3fb295a46d57" (UID: "25e07c34-72be-4c3a-8432-3fb295a46d57"). InnerVolumeSpecName "kube-api-access-bf8l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.106634 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-inventory" (OuterVolumeSpecName: "inventory") pod "25e07c34-72be-4c3a-8432-3fb295a46d57" (UID: "25e07c34-72be-4c3a-8432-3fb295a46d57"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.122346 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "25e07c34-72be-4c3a-8432-3fb295a46d57" (UID: "25e07c34-72be-4c3a-8432-3fb295a46d57"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.174340 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.174649 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf8l2\" (UniqueName: \"kubernetes.io/projected/25e07c34-72be-4c3a-8432-3fb295a46d57-kube-api-access-bf8l2\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.174660 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/25e07c34-72be-4c3a-8432-3fb295a46d57-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.466996 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" event={"ID":"25e07c34-72be-4c3a-8432-3fb295a46d57","Type":"ContainerDied","Data":"e706ece4b7e51d3d771b491e8fffbd4845ef5391bda1213c3861ebcde7ae57a4"} Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.467016 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kwkbw" Dec 04 14:18:54 crc kubenswrapper[4848]: I1204 14:18:54.467037 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e706ece4b7e51d3d771b491e8fffbd4845ef5391bda1213c3861ebcde7ae57a4" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.109094 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5"] Dec 04 14:18:55 crc kubenswrapper[4848]: E1204 14:18:55.111514 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e07c34-72be-4c3a-8432-3fb295a46d57" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.111848 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e07c34-72be-4c3a-8432-3fb295a46d57" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.112220 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e07c34-72be-4c3a-8432-3fb295a46d57" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.113201 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.115826 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.116871 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.116982 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.117374 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.137969 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5"] Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.302528 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.302620 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.302843 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.302877 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjdn7\" (UniqueName: \"kubernetes.io/projected/17b7cb97-4979-468f-89be-d14529b93be8-kube-api-access-wjdn7\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.404858 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.404910 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjdn7\" (UniqueName: \"kubernetes.io/projected/17b7cb97-4979-468f-89be-d14529b93be8-kube-api-access-wjdn7\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.404999 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.405044 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.410802 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.411524 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.418972 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.425047 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjdn7\" (UniqueName: \"kubernetes.io/projected/17b7cb97-4979-468f-89be-d14529b93be8-kube-api-access-wjdn7\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:55 crc kubenswrapper[4848]: I1204 14:18:55.437401 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:18:56 crc kubenswrapper[4848]: I1204 14:18:56.060507 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5"] Dec 04 14:18:56 crc kubenswrapper[4848]: W1204 14:18:56.062247 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17b7cb97_4979_468f_89be_d14529b93be8.slice/crio-759feaa9f8714b81dea36aaeb8b86e15568bf8bcf28d00b22318fd30eb86f5c7 WatchSource:0}: Error finding container 759feaa9f8714b81dea36aaeb8b86e15568bf8bcf28d00b22318fd30eb86f5c7: Status 404 returned error can't find the container with id 759feaa9f8714b81dea36aaeb8b86e15568bf8bcf28d00b22318fd30eb86f5c7 Dec 04 14:18:56 crc kubenswrapper[4848]: I1204 14:18:56.489620 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" event={"ID":"17b7cb97-4979-468f-89be-d14529b93be8","Type":"ContainerStarted","Data":"759feaa9f8714b81dea36aaeb8b86e15568bf8bcf28d00b22318fd30eb86f5c7"} Dec 04 14:18:57 crc kubenswrapper[4848]: I1204 14:18:57.504491 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" event={"ID":"17b7cb97-4979-468f-89be-d14529b93be8","Type":"ContainerStarted","Data":"3e4f8f69d447570f0e05ed0a765075a95787c8e955dcbc6ce2d4b5cde66291dc"} Dec 04 14:18:57 crc kubenswrapper[4848]: I1204 14:18:57.542297 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" podStartSLOduration=2.101559644 podStartE2EDuration="2.542269935s" podCreationTimestamp="2025-12-04 14:18:55 +0000 UTC" firstStartedPulling="2025-12-04 14:18:56.06701444 +0000 UTC m=+1840.009510968" lastFinishedPulling="2025-12-04 14:18:56.507724731 +0000 UTC m=+1840.450221259" observedRunningTime="2025-12-04 14:18:57.531747899 +0000 UTC m=+1841.474244437" watchObservedRunningTime="2025-12-04 14:18:57.542269935 +0000 UTC m=+1841.484766483" Dec 04 14:18:58 crc kubenswrapper[4848]: I1204 14:18:58.394112 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:18:58 crc kubenswrapper[4848]: E1204 14:18:58.394404 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:19:13 crc kubenswrapper[4848]: I1204 14:19:13.393662 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:19:13 crc kubenswrapper[4848]: E1204 14:19:13.394483 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:19:26 crc kubenswrapper[4848]: I1204 14:19:26.403357 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:19:26 crc kubenswrapper[4848]: E1204 14:19:26.404364 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:19:34 crc kubenswrapper[4848]: I1204 14:19:34.474054 4848 scope.go:117] "RemoveContainer" containerID="0754a58093d928184ce8017dad3911af3c33cc4056ba464d5fbecdcf2fc4dbb2" Dec 04 14:19:34 crc kubenswrapper[4848]: I1204 14:19:34.512778 4848 scope.go:117] "RemoveContainer" containerID="615184e23c8aeb0ec7d6d6030f4b2dcd7cf611c05c94c19a1cf1b3569d831af2" Dec 04 14:19:34 crc kubenswrapper[4848]: I1204 14:19:34.559031 4848 scope.go:117] "RemoveContainer" containerID="841d00b71ea982d6c3ddac488a7d77982f117bb6214829ecfacc33e54fa5011c" Dec 04 14:19:38 crc kubenswrapper[4848]: I1204 14:19:38.394266 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:19:38 crc kubenswrapper[4848]: E1204 14:19:38.395019 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:19:53 crc kubenswrapper[4848]: I1204 14:19:53.394541 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:19:53 crc kubenswrapper[4848]: E1204 14:19:53.395684 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:20:08 crc kubenswrapper[4848]: I1204 14:20:08.393988 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:20:08 crc kubenswrapper[4848]: E1204 14:20:08.394681 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:20:23 crc kubenswrapper[4848]: I1204 14:20:23.394198 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:20:23 crc kubenswrapper[4848]: I1204 14:20:23.702070 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"6068840f9ec731a3fcec181636bc276aedf6fad5ca2e8f03411d734389322f62"} Dec 04 14:20:48 crc kubenswrapper[4848]: I1204 14:20:48.446149 4848 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7ddb79586-x6vmg container/manager namespace/openshift-operators-redhat: Liveness probe status=failure output="Get \"http://10.217.0.49:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 14:20:48 crc kubenswrapper[4848]: I1204 14:20:48.447719 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators-redhat/loki-operator-controller-manager-7ddb79586-x6vmg" podUID="461d1ece-26ff-40ed-a92c-f21300adeec7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.49:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.378614 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g2d2n"] Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.382420 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.428101 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g84wl\" (UniqueName: \"kubernetes.io/projected/516c5c74-d871-4a8f-b81d-743930413cfc-kube-api-access-g84wl\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.428645 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-catalog-content\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.428751 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-utilities\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.436745 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g2d2n"] Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.531245 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-catalog-content\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.531324 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-utilities\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.531468 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g84wl\" (UniqueName: \"kubernetes.io/projected/516c5c74-d871-4a8f-b81d-743930413cfc-kube-api-access-g84wl\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.532292 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-catalog-content\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.532566 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-utilities\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.548995 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g84wl\" (UniqueName: \"kubernetes.io/projected/516c5c74-d871-4a8f-b81d-743930413cfc-kube-api-access-g84wl\") pod \"community-operators-g2d2n\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:04 crc kubenswrapper[4848]: I1204 14:21:04.712162 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:05 crc kubenswrapper[4848]: I1204 14:21:05.067352 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2z5jg"] Dec 04 14:21:05 crc kubenswrapper[4848]: I1204 14:21:05.081064 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2z5jg"] Dec 04 14:21:05 crc kubenswrapper[4848]: W1204 14:21:05.207126 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod516c5c74_d871_4a8f_b81d_743930413cfc.slice/crio-0c68039132b3141d212be1f7b63e5c5f85931af7a0bbc5e467e4894a103e09ac WatchSource:0}: Error finding container 0c68039132b3141d212be1f7b63e5c5f85931af7a0bbc5e467e4894a103e09ac: Status 404 returned error can't find the container with id 0c68039132b3141d212be1f7b63e5c5f85931af7a0bbc5e467e4894a103e09ac Dec 04 14:21:05 crc kubenswrapper[4848]: I1204 14:21:05.210975 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g2d2n"] Dec 04 14:21:06 crc kubenswrapper[4848]: I1204 14:21:06.225304 4848 generic.go:334] "Generic (PLEG): container finished" podID="516c5c74-d871-4a8f-b81d-743930413cfc" containerID="09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b" exitCode=0 Dec 04 14:21:06 crc kubenswrapper[4848]: I1204 14:21:06.225354 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g2d2n" event={"ID":"516c5c74-d871-4a8f-b81d-743930413cfc","Type":"ContainerDied","Data":"09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b"} Dec 04 14:21:06 crc kubenswrapper[4848]: I1204 14:21:06.225573 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g2d2n" event={"ID":"516c5c74-d871-4a8f-b81d-743930413cfc","Type":"ContainerStarted","Data":"0c68039132b3141d212be1f7b63e5c5f85931af7a0bbc5e467e4894a103e09ac"} Dec 04 14:21:06 crc kubenswrapper[4848]: I1204 14:21:06.227580 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:21:06 crc kubenswrapper[4848]: I1204 14:21:06.406143 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a530fe44-72a5-487a-8b28-fae47fa307c5" path="/var/lib/kubelet/pods/a530fe44-72a5-487a-8b28-fae47fa307c5/volumes" Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.040360 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-qbr2v"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.059504 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-qbr2v"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.079754 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-cgjg9"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.093604 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-mk75p"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.106941 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-47f3-account-create-update-sk9vk"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.119162 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-cgjg9"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.130806 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-99d5-account-create-update-j5z7r"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.146182 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-mk75p"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.161857 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-47f3-account-create-update-sk9vk"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.175908 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-99d5-account-create-update-j5z7r"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.189929 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-0e75-account-create-update-x7wmg"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.202581 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-0e75-account-create-update-x7wmg"] Dec 04 14:21:07 crc kubenswrapper[4848]: I1204 14:21:07.238258 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g2d2n" event={"ID":"516c5c74-d871-4a8f-b81d-743930413cfc","Type":"ContainerStarted","Data":"f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09"} Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.197908 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-fd6f-account-create-update-wwnqm"] Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.208705 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-fd6f-account-create-update-wwnqm"] Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.407406 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fa26501-8eec-4719-9b4d-211dc1e24ec9" path="/var/lib/kubelet/pods/2fa26501-8eec-4719-9b4d-211dc1e24ec9/volumes" Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.408630 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fa01532-be1d-4542-bad2-3608980a8835" path="/var/lib/kubelet/pods/6fa01532-be1d-4542-bad2-3608980a8835/volumes" Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.409382 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e78126d-0db0-4fee-a1b6-c23818909b5a" path="/var/lib/kubelet/pods/7e78126d-0db0-4fee-a1b6-c23818909b5a/volumes" Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.410095 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5592df9-94ea-4dd6-9f76-473b02a44897" path="/var/lib/kubelet/pods/b5592df9-94ea-4dd6-9f76-473b02a44897/volumes" Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.411306 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb9c5d54-6432-4f78-a4a1-9dc43ffff102" path="/var/lib/kubelet/pods/bb9c5d54-6432-4f78-a4a1-9dc43ffff102/volumes" Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.412550 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6dcffe4-90e2-4334-88c9-60b336d87402" path="/var/lib/kubelet/pods/d6dcffe4-90e2-4334-88c9-60b336d87402/volumes" Dec 04 14:21:08 crc kubenswrapper[4848]: I1204 14:21:08.413488 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed0bfa2e-a25b-4dce-86be-1401ac97f10a" path="/var/lib/kubelet/pods/ed0bfa2e-a25b-4dce-86be-1401ac97f10a/volumes" Dec 04 14:21:09 crc kubenswrapper[4848]: I1204 14:21:09.266849 4848 generic.go:334] "Generic (PLEG): container finished" podID="516c5c74-d871-4a8f-b81d-743930413cfc" containerID="f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09" exitCode=0 Dec 04 14:21:09 crc kubenswrapper[4848]: I1204 14:21:09.266901 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g2d2n" event={"ID":"516c5c74-d871-4a8f-b81d-743930413cfc","Type":"ContainerDied","Data":"f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09"} Dec 04 14:21:10 crc kubenswrapper[4848]: I1204 14:21:10.287012 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g2d2n" event={"ID":"516c5c74-d871-4a8f-b81d-743930413cfc","Type":"ContainerStarted","Data":"958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f"} Dec 04 14:21:10 crc kubenswrapper[4848]: I1204 14:21:10.324370 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g2d2n" podStartSLOduration=2.910521902 podStartE2EDuration="6.324339754s" podCreationTimestamp="2025-12-04 14:21:04 +0000 UTC" firstStartedPulling="2025-12-04 14:21:06.227368668 +0000 UTC m=+1970.169865196" lastFinishedPulling="2025-12-04 14:21:09.64118651 +0000 UTC m=+1973.583683048" observedRunningTime="2025-12-04 14:21:10.313516021 +0000 UTC m=+1974.256012549" watchObservedRunningTime="2025-12-04 14:21:10.324339754 +0000 UTC m=+1974.266836312" Dec 04 14:21:14 crc kubenswrapper[4848]: I1204 14:21:14.713648 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:14 crc kubenswrapper[4848]: I1204 14:21:14.715566 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:14 crc kubenswrapper[4848]: I1204 14:21:14.782241 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:15 crc kubenswrapper[4848]: I1204 14:21:15.409108 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:15 crc kubenswrapper[4848]: I1204 14:21:15.470183 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g2d2n"] Dec 04 14:21:17 crc kubenswrapper[4848]: I1204 14:21:17.044172 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4"] Dec 04 14:21:17 crc kubenswrapper[4848]: I1204 14:21:17.055747 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-bxkz4"] Dec 04 14:21:17 crc kubenswrapper[4848]: I1204 14:21:17.373920 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g2d2n" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" containerName="registry-server" containerID="cri-o://958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f" gracePeriod=2 Dec 04 14:21:17 crc kubenswrapper[4848]: I1204 14:21:17.936461 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.037526 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-7f5b-account-create-update-9j2qq"] Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.049406 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-7f5b-account-create-update-9j2qq"] Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.063927 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g84wl\" (UniqueName: \"kubernetes.io/projected/516c5c74-d871-4a8f-b81d-743930413cfc-kube-api-access-g84wl\") pod \"516c5c74-d871-4a8f-b81d-743930413cfc\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.064092 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-catalog-content\") pod \"516c5c74-d871-4a8f-b81d-743930413cfc\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.064189 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-utilities\") pod \"516c5c74-d871-4a8f-b81d-743930413cfc\" (UID: \"516c5c74-d871-4a8f-b81d-743930413cfc\") " Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.064852 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-utilities" (OuterVolumeSpecName: "utilities") pod "516c5c74-d871-4a8f-b81d-743930413cfc" (UID: "516c5c74-d871-4a8f-b81d-743930413cfc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.065211 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.070211 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/516c5c74-d871-4a8f-b81d-743930413cfc-kube-api-access-g84wl" (OuterVolumeSpecName: "kube-api-access-g84wl") pod "516c5c74-d871-4a8f-b81d-743930413cfc" (UID: "516c5c74-d871-4a8f-b81d-743930413cfc"). InnerVolumeSpecName "kube-api-access-g84wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.115272 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "516c5c74-d871-4a8f-b81d-743930413cfc" (UID: "516c5c74-d871-4a8f-b81d-743930413cfc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.168073 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g84wl\" (UniqueName: \"kubernetes.io/projected/516c5c74-d871-4a8f-b81d-743930413cfc-kube-api-access-g84wl\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.168103 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/516c5c74-d871-4a8f-b81d-743930413cfc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.386204 4848 generic.go:334] "Generic (PLEG): container finished" podID="516c5c74-d871-4a8f-b81d-743930413cfc" containerID="958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f" exitCode=0 Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.386259 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g2d2n" event={"ID":"516c5c74-d871-4a8f-b81d-743930413cfc","Type":"ContainerDied","Data":"958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f"} Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.386310 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g2d2n" event={"ID":"516c5c74-d871-4a8f-b81d-743930413cfc","Type":"ContainerDied","Data":"0c68039132b3141d212be1f7b63e5c5f85931af7a0bbc5e467e4894a103e09ac"} Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.386333 4848 scope.go:117] "RemoveContainer" containerID="958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.386412 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g2d2n" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.420858 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d945e4-9311-4d09-a75a-d6d1b5e59dff" path="/var/lib/kubelet/pods/48d945e4-9311-4d09-a75a-d6d1b5e59dff/volumes" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.422337 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e197c4da-c849-4c84-becb-dde8501eda45" path="/var/lib/kubelet/pods/e197c4da-c849-4c84-becb-dde8501eda45/volumes" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.427883 4848 scope.go:117] "RemoveContainer" containerID="f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.437303 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g2d2n"] Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.454278 4848 scope.go:117] "RemoveContainer" containerID="09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.455638 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g2d2n"] Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.508680 4848 scope.go:117] "RemoveContainer" containerID="958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f" Dec 04 14:21:18 crc kubenswrapper[4848]: E1204 14:21:18.509196 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f\": container with ID starting with 958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f not found: ID does not exist" containerID="958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.509278 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f"} err="failed to get container status \"958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f\": rpc error: code = NotFound desc = could not find container \"958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f\": container with ID starting with 958268871f8670174b4bffc43e871380540ee2981981273bafa251a9e597863f not found: ID does not exist" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.509306 4848 scope.go:117] "RemoveContainer" containerID="f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09" Dec 04 14:21:18 crc kubenswrapper[4848]: E1204 14:21:18.511175 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09\": container with ID starting with f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09 not found: ID does not exist" containerID="f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.511240 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09"} err="failed to get container status \"f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09\": rpc error: code = NotFound desc = could not find container \"f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09\": container with ID starting with f015f3d2103ab4fcfe34410855c091cf2b39bab1f9c61c1390b81f9cbbe9ad09 not found: ID does not exist" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.511279 4848 scope.go:117] "RemoveContainer" containerID="09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b" Dec 04 14:21:18 crc kubenswrapper[4848]: E1204 14:21:18.511645 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b\": container with ID starting with 09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b not found: ID does not exist" containerID="09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b" Dec 04 14:21:18 crc kubenswrapper[4848]: I1204 14:21:18.511673 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b"} err="failed to get container status \"09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b\": rpc error: code = NotFound desc = could not find container \"09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b\": container with ID starting with 09fac6b13e75e88ea468a32878bd8b7fbf17e4c2941249b0eebd54fd281b858b not found: ID does not exist" Dec 04 14:21:18 crc kubenswrapper[4848]: E1204 14:21:18.518048 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod516c5c74_d871_4a8f_b81d_743930413cfc.slice/crio-0c68039132b3141d212be1f7b63e5c5f85931af7a0bbc5e467e4894a103e09ac\": RecentStats: unable to find data in memory cache]" Dec 04 14:21:20 crc kubenswrapper[4848]: I1204 14:21:20.408538 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" path="/var/lib/kubelet/pods/516c5c74-d871-4a8f-b81d-743930413cfc/volumes" Dec 04 14:21:34 crc kubenswrapper[4848]: I1204 14:21:34.698171 4848 scope.go:117] "RemoveContainer" containerID="93a0f28489a9a7797f4f30b900479579fb3e7642d7dc0600f93932c052ddb1b3" Dec 04 14:21:34 crc kubenswrapper[4848]: I1204 14:21:34.734131 4848 scope.go:117] "RemoveContainer" containerID="af28b07da149fcebf8836123adc7855c4bfeff25ee48c659b5044cbd684ad376" Dec 04 14:21:34 crc kubenswrapper[4848]: I1204 14:21:34.805047 4848 scope.go:117] "RemoveContainer" containerID="a2cbfeeeaeff9ac2880d140d734bc5830537af1f4d7bfccbf22297c6650b0925" Dec 04 14:21:34 crc kubenswrapper[4848]: I1204 14:21:34.844703 4848 scope.go:117] "RemoveContainer" containerID="544f5b38aa92e3217e058b7d9623df4cec055b4aa04ae127917b6d64e77ac8ae" Dec 04 14:21:34 crc kubenswrapper[4848]: I1204 14:21:34.895810 4848 scope.go:117] "RemoveContainer" containerID="72a027a9649a4c0f6fc55e4d5cb03312847100751315821ac11a94d579d2652d" Dec 04 14:21:34 crc kubenswrapper[4848]: I1204 14:21:34.933747 4848 scope.go:117] "RemoveContainer" containerID="33a63e61f1aa5d7c9f1803584c7c970dfb19df092c2867dc4f89ecdff5877778" Dec 04 14:21:34 crc kubenswrapper[4848]: I1204 14:21:34.996996 4848 scope.go:117] "RemoveContainer" containerID="9097b4a5e0afac0434bf3a3eb2919f2f21b0524f320a1334550aaee2873df5af" Dec 04 14:21:35 crc kubenswrapper[4848]: I1204 14:21:35.017448 4848 scope.go:117] "RemoveContainer" containerID="34ef3765492d8cb8d0596a210dc5055bb9e57f155788c6e279aa6c369ae7712c" Dec 04 14:21:35 crc kubenswrapper[4848]: I1204 14:21:35.039343 4848 scope.go:117] "RemoveContainer" containerID="78d3f2e2146a537afba195dca6ac4e0413f68486d7cbd5ac0bb941406bfdff24" Dec 04 14:21:35 crc kubenswrapper[4848]: I1204 14:21:35.058257 4848 scope.go:117] "RemoveContainer" containerID="cd7bacadcb8cda904706d8879ccf0127bc1caf24053c6067a73e9cf2865a670b" Dec 04 14:21:35 crc kubenswrapper[4848]: I1204 14:21:35.083047 4848 scope.go:117] "RemoveContainer" containerID="777c333b317427bc294a1a4ce5830cf894f0f197a66027b0b4c274b2dc9d7ff9" Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.047093 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-021e-account-create-update-g264j"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.063959 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-dchv8"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.075552 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-fd58-account-create-update-r8csb"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.089376 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-61ea-account-create-update-mtrsm"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.099562 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-s2z5f"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.110109 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-021e-account-create-update-g264j"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.120058 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-dchv8"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.131185 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-fd58-account-create-update-r8csb"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.142543 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-61ea-account-create-update-mtrsm"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.154188 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-s2z5f"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.165742 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-ldg2g"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.176015 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-ldg2g"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.187776 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-2804-account-create-update-mrhhd"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.198081 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-2804-account-create-update-mrhhd"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.209534 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-cxgrr"] Dec 04 14:21:37 crc kubenswrapper[4848]: I1204 14:21:37.222740 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-cxgrr"] Dec 04 14:21:38 crc kubenswrapper[4848]: I1204 14:21:38.410050 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41c0df85-a096-4e29-9dae-4e98007e43ff" path="/var/lib/kubelet/pods/41c0df85-a096-4e29-9dae-4e98007e43ff/volumes" Dec 04 14:21:38 crc kubenswrapper[4848]: I1204 14:21:38.412407 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e722893-4a07-44e8-9e5a-3e04ff50dc47" path="/var/lib/kubelet/pods/4e722893-4a07-44e8-9e5a-3e04ff50dc47/volumes" Dec 04 14:21:38 crc kubenswrapper[4848]: I1204 14:21:38.413875 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a756447-5ce8-4b33-9075-e3b6100d807c" path="/var/lib/kubelet/pods/6a756447-5ce8-4b33-9075-e3b6100d807c/volumes" Dec 04 14:21:38 crc kubenswrapper[4848]: I1204 14:21:38.414603 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="867de14d-11f1-4306-873d-cdc3f6f90216" path="/var/lib/kubelet/pods/867de14d-11f1-4306-873d-cdc3f6f90216/volumes" Dec 04 14:21:38 crc kubenswrapper[4848]: I1204 14:21:38.416259 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f" path="/var/lib/kubelet/pods/97f4a8ad-b3cd-42ce-a7d2-78476b2aa16f/volumes" Dec 04 14:21:38 crc kubenswrapper[4848]: I1204 14:21:38.417061 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec" path="/var/lib/kubelet/pods/a1d785c9-6be8-49a5-a4f6-b20f8c93b1ec/volumes" Dec 04 14:21:38 crc kubenswrapper[4848]: I1204 14:21:38.418387 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cac2310b-905a-42df-92d8-611fa944c8f6" path="/var/lib/kubelet/pods/cac2310b-905a-42df-92d8-611fa944c8f6/volumes" Dec 04 14:21:38 crc kubenswrapper[4848]: I1204 14:21:38.419600 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df74ef9c-2afd-4236-901e-82c48594f70f" path="/var/lib/kubelet/pods/df74ef9c-2afd-4236-901e-82c48594f70f/volumes" Dec 04 14:21:48 crc kubenswrapper[4848]: I1204 14:21:48.048982 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-8pzdg"] Dec 04 14:21:48 crc kubenswrapper[4848]: I1204 14:21:48.059046 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-8pzdg"] Dec 04 14:21:48 crc kubenswrapper[4848]: I1204 14:21:48.407683 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf2c3a14-5cfd-49a1-9337-fb0b61611669" path="/var/lib/kubelet/pods/cf2c3a14-5cfd-49a1-9337-fb0b61611669/volumes" Dec 04 14:21:51 crc kubenswrapper[4848]: I1204 14:21:51.034839 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-r7w8t"] Dec 04 14:21:51 crc kubenswrapper[4848]: I1204 14:21:51.045988 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-r7w8t"] Dec 04 14:21:52 crc kubenswrapper[4848]: I1204 14:21:52.427051 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9925991d-d748-4a39-96a0-3bf2e391cad7" path="/var/lib/kubelet/pods/9925991d-d748-4a39-96a0-3bf2e391cad7/volumes" Dec 04 14:22:17 crc kubenswrapper[4848]: I1204 14:22:17.094069 4848 generic.go:334] "Generic (PLEG): container finished" podID="17b7cb97-4979-468f-89be-d14529b93be8" containerID="3e4f8f69d447570f0e05ed0a765075a95787c8e955dcbc6ce2d4b5cde66291dc" exitCode=0 Dec 04 14:22:17 crc kubenswrapper[4848]: I1204 14:22:17.094208 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" event={"ID":"17b7cb97-4979-468f-89be-d14529b93be8","Type":"ContainerDied","Data":"3e4f8f69d447570f0e05ed0a765075a95787c8e955dcbc6ce2d4b5cde66291dc"} Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.578498 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.719482 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjdn7\" (UniqueName: \"kubernetes.io/projected/17b7cb97-4979-468f-89be-d14529b93be8-kube-api-access-wjdn7\") pod \"17b7cb97-4979-468f-89be-d14529b93be8\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.719563 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-bootstrap-combined-ca-bundle\") pod \"17b7cb97-4979-468f-89be-d14529b93be8\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.719634 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-inventory\") pod \"17b7cb97-4979-468f-89be-d14529b93be8\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.719750 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-ssh-key\") pod \"17b7cb97-4979-468f-89be-d14529b93be8\" (UID: \"17b7cb97-4979-468f-89be-d14529b93be8\") " Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.725192 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "17b7cb97-4979-468f-89be-d14529b93be8" (UID: "17b7cb97-4979-468f-89be-d14529b93be8"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.727012 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17b7cb97-4979-468f-89be-d14529b93be8-kube-api-access-wjdn7" (OuterVolumeSpecName: "kube-api-access-wjdn7") pod "17b7cb97-4979-468f-89be-d14529b93be8" (UID: "17b7cb97-4979-468f-89be-d14529b93be8"). InnerVolumeSpecName "kube-api-access-wjdn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.754666 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-inventory" (OuterVolumeSpecName: "inventory") pod "17b7cb97-4979-468f-89be-d14529b93be8" (UID: "17b7cb97-4979-468f-89be-d14529b93be8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.756215 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "17b7cb97-4979-468f-89be-d14529b93be8" (UID: "17b7cb97-4979-468f-89be-d14529b93be8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.823930 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjdn7\" (UniqueName: \"kubernetes.io/projected/17b7cb97-4979-468f-89be-d14529b93be8-kube-api-access-wjdn7\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.823985 4848 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.823996 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:18 crc kubenswrapper[4848]: I1204 14:22:18.824005 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17b7cb97-4979-468f-89be-d14529b93be8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.117936 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" event={"ID":"17b7cb97-4979-468f-89be-d14529b93be8","Type":"ContainerDied","Data":"759feaa9f8714b81dea36aaeb8b86e15568bf8bcf28d00b22318fd30eb86f5c7"} Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.118289 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="759feaa9f8714b81dea36aaeb8b86e15568bf8bcf28d00b22318fd30eb86f5c7" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.118034 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.213961 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp"] Dec 04 14:22:19 crc kubenswrapper[4848]: E1204 14:22:19.214425 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" containerName="registry-server" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.214442 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" containerName="registry-server" Dec 04 14:22:19 crc kubenswrapper[4848]: E1204 14:22:19.214467 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" containerName="extract-content" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.214474 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" containerName="extract-content" Dec 04 14:22:19 crc kubenswrapper[4848]: E1204 14:22:19.214505 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17b7cb97-4979-468f-89be-d14529b93be8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.214514 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="17b7cb97-4979-468f-89be-d14529b93be8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 14:22:19 crc kubenswrapper[4848]: E1204 14:22:19.214524 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" containerName="extract-utilities" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.214530 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" containerName="extract-utilities" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.214738 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="516c5c74-d871-4a8f-b81d-743930413cfc" containerName="registry-server" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.214757 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="17b7cb97-4979-468f-89be-d14529b93be8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.215575 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.224924 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.225235 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.225534 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.225695 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.232370 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp"] Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.347895 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.347997 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5ngm\" (UniqueName: \"kubernetes.io/projected/9d69a925-9c1f-4405-9300-e6c4c3cada23-kube-api-access-s5ngm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.348295 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.452479 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.453073 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.453245 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5ngm\" (UniqueName: \"kubernetes.io/projected/9d69a925-9c1f-4405-9300-e6c4c3cada23-kube-api-access-s5ngm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.458517 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.471559 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.475460 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5ngm\" (UniqueName: \"kubernetes.io/projected/9d69a925-9c1f-4405-9300-e6c4c3cada23-kube-api-access-s5ngm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:19 crc kubenswrapper[4848]: I1204 14:22:19.536619 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:22:20 crc kubenswrapper[4848]: I1204 14:22:20.092519 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp"] Dec 04 14:22:20 crc kubenswrapper[4848]: W1204 14:22:20.098598 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d69a925_9c1f_4405_9300_e6c4c3cada23.slice/crio-b39de2be5fe8ec92c1a4e9c3023715bf1128fe0ad1964d27789e0a28c7ab7ea4 WatchSource:0}: Error finding container b39de2be5fe8ec92c1a4e9c3023715bf1128fe0ad1964d27789e0a28c7ab7ea4: Status 404 returned error can't find the container with id b39de2be5fe8ec92c1a4e9c3023715bf1128fe0ad1964d27789e0a28c7ab7ea4 Dec 04 14:22:20 crc kubenswrapper[4848]: I1204 14:22:20.130531 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" event={"ID":"9d69a925-9c1f-4405-9300-e6c4c3cada23","Type":"ContainerStarted","Data":"b39de2be5fe8ec92c1a4e9c3023715bf1128fe0ad1964d27789e0a28c7ab7ea4"} Dec 04 14:22:21 crc kubenswrapper[4848]: I1204 14:22:21.151596 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" event={"ID":"9d69a925-9c1f-4405-9300-e6c4c3cada23","Type":"ContainerStarted","Data":"aadc4b5b043c1c9343fd04fde8366c5c837d2ce9c2dce94b1c64be0e3edfd285"} Dec 04 14:22:21 crc kubenswrapper[4848]: I1204 14:22:21.182583 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" podStartSLOduration=1.6538568439999999 podStartE2EDuration="2.18255753s" podCreationTimestamp="2025-12-04 14:22:19 +0000 UTC" firstStartedPulling="2025-12-04 14:22:20.10218369 +0000 UTC m=+2044.044680208" lastFinishedPulling="2025-12-04 14:22:20.630884366 +0000 UTC m=+2044.573380894" observedRunningTime="2025-12-04 14:22:21.171131261 +0000 UTC m=+2045.113627789" watchObservedRunningTime="2025-12-04 14:22:21.18255753 +0000 UTC m=+2045.125054078" Dec 04 14:22:26 crc kubenswrapper[4848]: I1204 14:22:26.048424 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-6hkhx"] Dec 04 14:22:26 crc kubenswrapper[4848]: I1204 14:22:26.059086 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-6hkhx"] Dec 04 14:22:26 crc kubenswrapper[4848]: I1204 14:22:26.405147 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f64ca832-b3c0-45a9-b967-a8f3ce9af156" path="/var/lib/kubelet/pods/f64ca832-b3c0-45a9-b967-a8f3ce9af156/volumes" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.390187 4848 scope.go:117] "RemoveContainer" containerID="12c88f7d42b296bb37e58011445ef99d0f1697bb57702df574b9329bcbdf2f39" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.438098 4848 scope.go:117] "RemoveContainer" containerID="8283fa960e8727cef37f7a575c8d78f76e0df4c960df766a40c8f256345b71c7" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.488029 4848 scope.go:117] "RemoveContainer" containerID="5e9d418ccd6305d98dee61226d0bb4de9c1d0061f7bec7ab92a5ae7840ba2398" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.545554 4848 scope.go:117] "RemoveContainer" containerID="5b709be838cafa6f4d427b1f3e0a3be519e9629525ded45e7ff845d623dd9a82" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.590857 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lhdrn"] Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.593795 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.604531 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhdrn"] Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.627967 4848 scope.go:117] "RemoveContainer" containerID="391a991f094c6871eb4b127ade42b2169669f19a1271921c0e6fd15574eca231" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.666393 4848 scope.go:117] "RemoveContainer" containerID="93f0d85a257f7eb16a75dec0e34adf2c4206b616299aaa2ff15080fde79aa7fc" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.711626 4848 scope.go:117] "RemoveContainer" containerID="1d2280872e954728b571252f7e9904fc32854a28826a1e941433fbeb18ae94ba" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.736790 4848 scope.go:117] "RemoveContainer" containerID="a68eff2e9b3822442aff1993cc6d0135e5d3bbac4dc7536797a5c55789a98e01" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.763174 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swp5c\" (UniqueName: \"kubernetes.io/projected/a90f9327-1292-4ea7-bc56-457bd4509ad3-kube-api-access-swp5c\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.763439 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-utilities\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.763859 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-catalog-content\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.774234 4848 scope.go:117] "RemoveContainer" containerID="2728ee54f8ea9c9f2e08c4ed603b6ddb8e56b9211c9c2f1a3db03090976d8cc9" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.798505 4848 scope.go:117] "RemoveContainer" containerID="2ded7473b5d4b3564f6fc47d069bdb2ecb6bd7939ef27e3dfe27f9848e205a6a" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.834148 4848 scope.go:117] "RemoveContainer" containerID="5bf60c5607da1252b532d01b4a4deee702442142969cf45cc40dc08341cee54e" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.866075 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-catalog-content\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.866184 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swp5c\" (UniqueName: \"kubernetes.io/projected/a90f9327-1292-4ea7-bc56-457bd4509ad3-kube-api-access-swp5c\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.866287 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-utilities\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.866726 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-catalog-content\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.866796 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-utilities\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:35 crc kubenswrapper[4848]: I1204 14:22:35.890724 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swp5c\" (UniqueName: \"kubernetes.io/projected/a90f9327-1292-4ea7-bc56-457bd4509ad3-kube-api-access-swp5c\") pod \"redhat-marketplace-lhdrn\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:36 crc kubenswrapper[4848]: I1204 14:22:36.003679 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:36 crc kubenswrapper[4848]: I1204 14:22:36.355242 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhdrn"] Dec 04 14:22:37 crc kubenswrapper[4848]: I1204 14:22:37.035390 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-t9r82"] Dec 04 14:22:37 crc kubenswrapper[4848]: I1204 14:22:37.051146 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-t9r82"] Dec 04 14:22:37 crc kubenswrapper[4848]: I1204 14:22:37.348695 4848 generic.go:334] "Generic (PLEG): container finished" podID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerID="91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904" exitCode=0 Dec 04 14:22:37 crc kubenswrapper[4848]: I1204 14:22:37.348739 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhdrn" event={"ID":"a90f9327-1292-4ea7-bc56-457bd4509ad3","Type":"ContainerDied","Data":"91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904"} Dec 04 14:22:37 crc kubenswrapper[4848]: I1204 14:22:37.348795 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhdrn" event={"ID":"a90f9327-1292-4ea7-bc56-457bd4509ad3","Type":"ContainerStarted","Data":"1293d46a7871fdc37f8c235f901586a9154e9c5782e53208db02c271edd15725"} Dec 04 14:22:38 crc kubenswrapper[4848]: I1204 14:22:38.372937 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhdrn" event={"ID":"a90f9327-1292-4ea7-bc56-457bd4509ad3","Type":"ContainerStarted","Data":"a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8"} Dec 04 14:22:38 crc kubenswrapper[4848]: I1204 14:22:38.415068 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deb8ffb1-2a45-4591-b749-ef7c9b2425bb" path="/var/lib/kubelet/pods/deb8ffb1-2a45-4591-b749-ef7c9b2425bb/volumes" Dec 04 14:22:39 crc kubenswrapper[4848]: I1204 14:22:39.042847 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-l5jbz"] Dec 04 14:22:39 crc kubenswrapper[4848]: I1204 14:22:39.063297 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-l5jbz"] Dec 04 14:22:39 crc kubenswrapper[4848]: I1204 14:22:39.383866 4848 generic.go:334] "Generic (PLEG): container finished" podID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerID="a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8" exitCode=0 Dec 04 14:22:39 crc kubenswrapper[4848]: I1204 14:22:39.383988 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhdrn" event={"ID":"a90f9327-1292-4ea7-bc56-457bd4509ad3","Type":"ContainerDied","Data":"a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8"} Dec 04 14:22:40 crc kubenswrapper[4848]: I1204 14:22:40.413347 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23eae77d-63ca-4004-969b-cfac163d3975" path="/var/lib/kubelet/pods/23eae77d-63ca-4004-969b-cfac163d3975/volumes" Dec 04 14:22:40 crc kubenswrapper[4848]: I1204 14:22:40.414454 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhdrn" event={"ID":"a90f9327-1292-4ea7-bc56-457bd4509ad3","Type":"ContainerStarted","Data":"4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108"} Dec 04 14:22:40 crc kubenswrapper[4848]: I1204 14:22:40.443833 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lhdrn" podStartSLOduration=3.033141315 podStartE2EDuration="5.443803531s" podCreationTimestamp="2025-12-04 14:22:35 +0000 UTC" firstStartedPulling="2025-12-04 14:22:37.350740423 +0000 UTC m=+2061.293236941" lastFinishedPulling="2025-12-04 14:22:39.761402629 +0000 UTC m=+2063.703899157" observedRunningTime="2025-12-04 14:22:40.425520557 +0000 UTC m=+2064.368017105" watchObservedRunningTime="2025-12-04 14:22:40.443803531 +0000 UTC m=+2064.386300069" Dec 04 14:22:44 crc kubenswrapper[4848]: I1204 14:22:44.314696 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:22:44 crc kubenswrapper[4848]: I1204 14:22:44.315283 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:22:46 crc kubenswrapper[4848]: I1204 14:22:46.003943 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:46 crc kubenswrapper[4848]: I1204 14:22:46.005224 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:46 crc kubenswrapper[4848]: I1204 14:22:46.063368 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-m9h4t"] Dec 04 14:22:46 crc kubenswrapper[4848]: I1204 14:22:46.075496 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-m9h4t"] Dec 04 14:22:46 crc kubenswrapper[4848]: I1204 14:22:46.093962 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:46 crc kubenswrapper[4848]: I1204 14:22:46.412351 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7380f023-2712-4a2b-a193-ec401d07c9ad" path="/var/lib/kubelet/pods/7380f023-2712-4a2b-a193-ec401d07c9ad/volumes" Dec 04 14:22:46 crc kubenswrapper[4848]: I1204 14:22:46.538350 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:46 crc kubenswrapper[4848]: I1204 14:22:46.598655 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhdrn"] Dec 04 14:22:48 crc kubenswrapper[4848]: I1204 14:22:48.503098 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lhdrn" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerName="registry-server" containerID="cri-o://4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108" gracePeriod=2 Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.084005 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.115940 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swp5c\" (UniqueName: \"kubernetes.io/projected/a90f9327-1292-4ea7-bc56-457bd4509ad3-kube-api-access-swp5c\") pod \"a90f9327-1292-4ea7-bc56-457bd4509ad3\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.116495 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-catalog-content\") pod \"a90f9327-1292-4ea7-bc56-457bd4509ad3\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.117800 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-utilities\") pod \"a90f9327-1292-4ea7-bc56-457bd4509ad3\" (UID: \"a90f9327-1292-4ea7-bc56-457bd4509ad3\") " Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.120197 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-utilities" (OuterVolumeSpecName: "utilities") pod "a90f9327-1292-4ea7-bc56-457bd4509ad3" (UID: "a90f9327-1292-4ea7-bc56-457bd4509ad3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.132808 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90f9327-1292-4ea7-bc56-457bd4509ad3-kube-api-access-swp5c" (OuterVolumeSpecName: "kube-api-access-swp5c") pod "a90f9327-1292-4ea7-bc56-457bd4509ad3" (UID: "a90f9327-1292-4ea7-bc56-457bd4509ad3"). InnerVolumeSpecName "kube-api-access-swp5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.137886 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a90f9327-1292-4ea7-bc56-457bd4509ad3" (UID: "a90f9327-1292-4ea7-bc56-457bd4509ad3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.221730 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.221779 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a90f9327-1292-4ea7-bc56-457bd4509ad3-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.221795 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swp5c\" (UniqueName: \"kubernetes.io/projected/a90f9327-1292-4ea7-bc56-457bd4509ad3-kube-api-access-swp5c\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.517426 4848 generic.go:334] "Generic (PLEG): container finished" podID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerID="4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108" exitCode=0 Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.517471 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhdrn" event={"ID":"a90f9327-1292-4ea7-bc56-457bd4509ad3","Type":"ContainerDied","Data":"4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108"} Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.517493 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhdrn" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.517512 4848 scope.go:117] "RemoveContainer" containerID="4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.517500 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhdrn" event={"ID":"a90f9327-1292-4ea7-bc56-457bd4509ad3","Type":"ContainerDied","Data":"1293d46a7871fdc37f8c235f901586a9154e9c5782e53208db02c271edd15725"} Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.561496 4848 scope.go:117] "RemoveContainer" containerID="a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.586854 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhdrn"] Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.608445 4848 scope.go:117] "RemoveContainer" containerID="91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.627966 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhdrn"] Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.678340 4848 scope.go:117] "RemoveContainer" containerID="4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108" Dec 04 14:22:49 crc kubenswrapper[4848]: E1204 14:22:49.678776 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108\": container with ID starting with 4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108 not found: ID does not exist" containerID="4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.678815 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108"} err="failed to get container status \"4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108\": rpc error: code = NotFound desc = could not find container \"4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108\": container with ID starting with 4185884350f852ee2eda69186fcb27900fc9f8ad6ce5c863381d70fb33a66108 not found: ID does not exist" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.678841 4848 scope.go:117] "RemoveContainer" containerID="a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8" Dec 04 14:22:49 crc kubenswrapper[4848]: E1204 14:22:49.679217 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8\": container with ID starting with a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8 not found: ID does not exist" containerID="a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.679263 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8"} err="failed to get container status \"a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8\": rpc error: code = NotFound desc = could not find container \"a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8\": container with ID starting with a6b8433eb54d6ab8c02c4e19993ee1c7de5d9703056842403c69abae51ba9ac8 not found: ID does not exist" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.679293 4848 scope.go:117] "RemoveContainer" containerID="91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904" Dec 04 14:22:49 crc kubenswrapper[4848]: E1204 14:22:49.679707 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904\": container with ID starting with 91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904 not found: ID does not exist" containerID="91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904" Dec 04 14:22:49 crc kubenswrapper[4848]: I1204 14:22:49.679737 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904"} err="failed to get container status \"91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904\": rpc error: code = NotFound desc = could not find container \"91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904\": container with ID starting with 91f5e5ff94fdcd18d282539053f480b7c1b84dd57346df0a9090489f286a6904 not found: ID does not exist" Dec 04 14:22:50 crc kubenswrapper[4848]: I1204 14:22:50.411252 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" path="/var/lib/kubelet/pods/a90f9327-1292-4ea7-bc56-457bd4509ad3/volumes" Dec 04 14:22:53 crc kubenswrapper[4848]: I1204 14:22:53.034828 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-4dpj6"] Dec 04 14:22:53 crc kubenswrapper[4848]: I1204 14:22:53.043993 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-4dpj6"] Dec 04 14:22:54 crc kubenswrapper[4848]: I1204 14:22:54.409775 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47da25b0-48bd-4dd1-8ad9-6db3f8565ad5" path="/var/lib/kubelet/pods/47da25b0-48bd-4dd1-8ad9-6db3f8565ad5/volumes" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.716301 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l7lqn"] Dec 04 14:23:10 crc kubenswrapper[4848]: E1204 14:23:10.717640 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerName="registry-server" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.717680 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerName="registry-server" Dec 04 14:23:10 crc kubenswrapper[4848]: E1204 14:23:10.717708 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerName="extract-utilities" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.717716 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerName="extract-utilities" Dec 04 14:23:10 crc kubenswrapper[4848]: E1204 14:23:10.717754 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerName="extract-content" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.717761 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerName="extract-content" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.717995 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a90f9327-1292-4ea7-bc56-457bd4509ad3" containerName="registry-server" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.719820 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.735040 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l7lqn"] Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.827919 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk427\" (UniqueName: \"kubernetes.io/projected/ae352946-5e42-4d28-87d0-389cea7430ae-kube-api-access-rk427\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.828398 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-utilities\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.828460 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-catalog-content\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.932318 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-utilities\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.932382 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-catalog-content\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.932578 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk427\" (UniqueName: \"kubernetes.io/projected/ae352946-5e42-4d28-87d0-389cea7430ae-kube-api-access-rk427\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.933516 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-utilities\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.933767 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-catalog-content\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:10 crc kubenswrapper[4848]: I1204 14:23:10.953386 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk427\" (UniqueName: \"kubernetes.io/projected/ae352946-5e42-4d28-87d0-389cea7430ae-kube-api-access-rk427\") pod \"redhat-operators-l7lqn\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:11 crc kubenswrapper[4848]: I1204 14:23:11.064167 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:11 crc kubenswrapper[4848]: I1204 14:23:11.606409 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l7lqn"] Dec 04 14:23:11 crc kubenswrapper[4848]: I1204 14:23:11.791509 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7lqn" event={"ID":"ae352946-5e42-4d28-87d0-389cea7430ae","Type":"ContainerStarted","Data":"6ac60d3bb873e2072a1646dcd38cc3917a97af8dc382c5f20da45ba1219ad267"} Dec 04 14:23:12 crc kubenswrapper[4848]: I1204 14:23:12.806458 4848 generic.go:334] "Generic (PLEG): container finished" podID="ae352946-5e42-4d28-87d0-389cea7430ae" containerID="94ad83ed9540698a5867f444cfcc665e5a078a9675e5e8bddf03532a5890cf79" exitCode=0 Dec 04 14:23:12 crc kubenswrapper[4848]: I1204 14:23:12.806584 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7lqn" event={"ID":"ae352946-5e42-4d28-87d0-389cea7430ae","Type":"ContainerDied","Data":"94ad83ed9540698a5867f444cfcc665e5a078a9675e5e8bddf03532a5890cf79"} Dec 04 14:23:13 crc kubenswrapper[4848]: I1204 14:23:13.818446 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7lqn" event={"ID":"ae352946-5e42-4d28-87d0-389cea7430ae","Type":"ContainerStarted","Data":"6e50c3861ce5696cf25abc45ef97f4bc151bc956144971daecfcd3748646deee"} Dec 04 14:23:14 crc kubenswrapper[4848]: I1204 14:23:14.314521 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:23:14 crc kubenswrapper[4848]: I1204 14:23:14.314583 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:23:17 crc kubenswrapper[4848]: I1204 14:23:17.859907 4848 generic.go:334] "Generic (PLEG): container finished" podID="ae352946-5e42-4d28-87d0-389cea7430ae" containerID="6e50c3861ce5696cf25abc45ef97f4bc151bc956144971daecfcd3748646deee" exitCode=0 Dec 04 14:23:17 crc kubenswrapper[4848]: I1204 14:23:17.860095 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7lqn" event={"ID":"ae352946-5e42-4d28-87d0-389cea7430ae","Type":"ContainerDied","Data":"6e50c3861ce5696cf25abc45ef97f4bc151bc956144971daecfcd3748646deee"} Dec 04 14:23:18 crc kubenswrapper[4848]: I1204 14:23:18.873633 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7lqn" event={"ID":"ae352946-5e42-4d28-87d0-389cea7430ae","Type":"ContainerStarted","Data":"36482d8e2eb383aa6c89a3f2c137fad7eb95b3137fb09e7ce36d7d2b6e7fef7a"} Dec 04 14:23:18 crc kubenswrapper[4848]: I1204 14:23:18.903195 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l7lqn" podStartSLOduration=3.417552157 podStartE2EDuration="8.903176731s" podCreationTimestamp="2025-12-04 14:23:10 +0000 UTC" firstStartedPulling="2025-12-04 14:23:12.808597881 +0000 UTC m=+2096.751094409" lastFinishedPulling="2025-12-04 14:23:18.294222455 +0000 UTC m=+2102.236718983" observedRunningTime="2025-12-04 14:23:18.901970893 +0000 UTC m=+2102.844467441" watchObservedRunningTime="2025-12-04 14:23:18.903176731 +0000 UTC m=+2102.845673259" Dec 04 14:23:21 crc kubenswrapper[4848]: I1204 14:23:21.064635 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:21 crc kubenswrapper[4848]: I1204 14:23:21.065216 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:22 crc kubenswrapper[4848]: I1204 14:23:22.115282 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l7lqn" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="registry-server" probeResult="failure" output=< Dec 04 14:23:22 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 14:23:22 crc kubenswrapper[4848]: > Dec 04 14:23:31 crc kubenswrapper[4848]: I1204 14:23:31.113249 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:31 crc kubenswrapper[4848]: I1204 14:23:31.179829 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:31 crc kubenswrapper[4848]: I1204 14:23:31.356404 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l7lqn"] Dec 04 14:23:33 crc kubenswrapper[4848]: I1204 14:23:33.023830 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l7lqn" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="registry-server" containerID="cri-o://36482d8e2eb383aa6c89a3f2c137fad7eb95b3137fb09e7ce36d7d2b6e7fef7a" gracePeriod=2 Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.065668 4848 generic.go:334] "Generic (PLEG): container finished" podID="ae352946-5e42-4d28-87d0-389cea7430ae" containerID="36482d8e2eb383aa6c89a3f2c137fad7eb95b3137fb09e7ce36d7d2b6e7fef7a" exitCode=0 Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.065704 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7lqn" event={"ID":"ae352946-5e42-4d28-87d0-389cea7430ae","Type":"ContainerDied","Data":"36482d8e2eb383aa6c89a3f2c137fad7eb95b3137fb09e7ce36d7d2b6e7fef7a"} Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.295201 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.413125 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-utilities\") pod \"ae352946-5e42-4d28-87d0-389cea7430ae\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.413597 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-catalog-content\") pod \"ae352946-5e42-4d28-87d0-389cea7430ae\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.413741 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk427\" (UniqueName: \"kubernetes.io/projected/ae352946-5e42-4d28-87d0-389cea7430ae-kube-api-access-rk427\") pod \"ae352946-5e42-4d28-87d0-389cea7430ae\" (UID: \"ae352946-5e42-4d28-87d0-389cea7430ae\") " Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.414655 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-utilities" (OuterVolumeSpecName: "utilities") pod "ae352946-5e42-4d28-87d0-389cea7430ae" (UID: "ae352946-5e42-4d28-87d0-389cea7430ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.422578 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae352946-5e42-4d28-87d0-389cea7430ae-kube-api-access-rk427" (OuterVolumeSpecName: "kube-api-access-rk427") pod "ae352946-5e42-4d28-87d0-389cea7430ae" (UID: "ae352946-5e42-4d28-87d0-389cea7430ae"). InnerVolumeSpecName "kube-api-access-rk427". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.504651 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae352946-5e42-4d28-87d0-389cea7430ae" (UID: "ae352946-5e42-4d28-87d0-389cea7430ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.516842 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.516869 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk427\" (UniqueName: \"kubernetes.io/projected/ae352946-5e42-4d28-87d0-389cea7430ae-kube-api-access-rk427\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:34 crc kubenswrapper[4848]: I1204 14:23:34.516881 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae352946-5e42-4d28-87d0-389cea7430ae-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:23:35 crc kubenswrapper[4848]: I1204 14:23:35.079568 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l7lqn" event={"ID":"ae352946-5e42-4d28-87d0-389cea7430ae","Type":"ContainerDied","Data":"6ac60d3bb873e2072a1646dcd38cc3917a97af8dc382c5f20da45ba1219ad267"} Dec 04 14:23:35 crc kubenswrapper[4848]: I1204 14:23:35.079682 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l7lqn" Dec 04 14:23:35 crc kubenswrapper[4848]: I1204 14:23:35.079968 4848 scope.go:117] "RemoveContainer" containerID="36482d8e2eb383aa6c89a3f2c137fad7eb95b3137fb09e7ce36d7d2b6e7fef7a" Dec 04 14:23:35 crc kubenswrapper[4848]: I1204 14:23:35.113441 4848 scope.go:117] "RemoveContainer" containerID="6e50c3861ce5696cf25abc45ef97f4bc151bc956144971daecfcd3748646deee" Dec 04 14:23:35 crc kubenswrapper[4848]: I1204 14:23:35.120211 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l7lqn"] Dec 04 14:23:35 crc kubenswrapper[4848]: I1204 14:23:35.134885 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l7lqn"] Dec 04 14:23:35 crc kubenswrapper[4848]: I1204 14:23:35.144529 4848 scope.go:117] "RemoveContainer" containerID="94ad83ed9540698a5867f444cfcc665e5a078a9675e5e8bddf03532a5890cf79" Dec 04 14:23:36 crc kubenswrapper[4848]: I1204 14:23:36.156354 4848 scope.go:117] "RemoveContainer" containerID="092b1f360bd7cd0348b2807080cbd29d7a7ba4fc860af7e4bb8a684d87b00a6e" Dec 04 14:23:36 crc kubenswrapper[4848]: I1204 14:23:36.187396 4848 scope.go:117] "RemoveContainer" containerID="570461ea30f63038f02b2b2ddb9a7ccd81b193ba9bfe54003f38f4e9c768ceb1" Dec 04 14:23:36 crc kubenswrapper[4848]: I1204 14:23:36.266347 4848 scope.go:117] "RemoveContainer" containerID="2f6081dfc3eaf61bd0b3a94ff130d966d45fb0be83fd461934d3e37ecd86f81c" Dec 04 14:23:36 crc kubenswrapper[4848]: I1204 14:23:36.324669 4848 scope.go:117] "RemoveContainer" containerID="f280ccc815c96ade43f1fc073378ab7a2f03d67964c55418b44f3d5c2300f082" Dec 04 14:23:36 crc kubenswrapper[4848]: I1204 14:23:36.410829 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" path="/var/lib/kubelet/pods/ae352946-5e42-4d28-87d0-389cea7430ae/volumes" Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.069504 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pk76z"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.082157 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-tn2wq"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.095114 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3873-account-create-update-nc5n8"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.111391 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-caff-account-create-update-sgpcw"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.127577 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-e824-account-create-update-h466m"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.138179 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-6fc5j"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.148370 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-caff-account-create-update-sgpcw"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.158291 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3873-account-create-update-nc5n8"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.170144 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pk76z"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.181178 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-tn2wq"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.191375 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-6fc5j"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.200732 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-e824-account-create-update-h466m"] Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.407965 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd79f24-e81e-49bd-917d-5a6e3281f336" path="/var/lib/kubelet/pods/2bd79f24-e81e-49bd-917d-5a6e3281f336/volumes" Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.409171 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34db090c-22b5-4462-a0ef-aca71a75a21a" path="/var/lib/kubelet/pods/34db090c-22b5-4462-a0ef-aca71a75a21a/volumes" Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.409971 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75e19b5f-83d0-4a02-8791-4da2749e0a12" path="/var/lib/kubelet/pods/75e19b5f-83d0-4a02-8791-4da2749e0a12/volumes" Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.410697 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d94a8c9d-5f73-44fa-8fc7-f725f21a8127" path="/var/lib/kubelet/pods/d94a8c9d-5f73-44fa-8fc7-f725f21a8127/volumes" Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.412086 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f692bc07-eca1-4b0b-b5fc-a9ad3aa34232" path="/var/lib/kubelet/pods/f692bc07-eca1-4b0b-b5fc-a9ad3aa34232/volumes" Dec 04 14:23:38 crc kubenswrapper[4848]: I1204 14:23:38.412985 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcc6ad68-0f83-4a73-b83a-46eca178604e" path="/var/lib/kubelet/pods/fcc6ad68-0f83-4a73-b83a-46eca178604e/volumes" Dec 04 14:23:44 crc kubenswrapper[4848]: I1204 14:23:44.313865 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:23:44 crc kubenswrapper[4848]: I1204 14:23:44.315089 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:23:44 crc kubenswrapper[4848]: I1204 14:23:44.315171 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:23:44 crc kubenswrapper[4848]: I1204 14:23:44.316412 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6068840f9ec731a3fcec181636bc276aedf6fad5ca2e8f03411d734389322f62"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:23:44 crc kubenswrapper[4848]: I1204 14:23:44.316538 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://6068840f9ec731a3fcec181636bc276aedf6fad5ca2e8f03411d734389322f62" gracePeriod=600 Dec 04 14:23:45 crc kubenswrapper[4848]: I1204 14:23:45.194848 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="6068840f9ec731a3fcec181636bc276aedf6fad5ca2e8f03411d734389322f62" exitCode=0 Dec 04 14:23:45 crc kubenswrapper[4848]: I1204 14:23:45.194943 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"6068840f9ec731a3fcec181636bc276aedf6fad5ca2e8f03411d734389322f62"} Dec 04 14:23:45 crc kubenswrapper[4848]: I1204 14:23:45.195142 4848 scope.go:117] "RemoveContainer" containerID="7c864c9e7985a9eb14ea5000b603960666660cb0f2ff7a3e907a048854c73357" Dec 04 14:23:46 crc kubenswrapper[4848]: I1204 14:23:46.208620 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6"} Dec 04 14:24:12 crc kubenswrapper[4848]: I1204 14:24:12.050283 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4czxk"] Dec 04 14:24:12 crc kubenswrapper[4848]: I1204 14:24:12.062035 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4czxk"] Dec 04 14:24:12 crc kubenswrapper[4848]: I1204 14:24:12.408244 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="013c9f72-2dcd-4e71-a4ab-aec59a8131aa" path="/var/lib/kubelet/pods/013c9f72-2dcd-4e71-a4ab-aec59a8131aa/volumes" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.051078 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-sl25v"] Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.061408 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-sl25v"] Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.414872 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db1d9746-4bfb-48fb-b965-bf89eb95885c" path="/var/lib/kubelet/pods/db1d9746-4bfb-48fb-b965-bf89eb95885c/volumes" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.526330 4848 scope.go:117] "RemoveContainer" containerID="24bd7e7b81eb628112fa037f245a03d4b2f307ab5c11fb2a7de3b4013b9afa74" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.558406 4848 scope.go:117] "RemoveContainer" containerID="10996e0c1d719def5deae568a099963cc051c51e42b330c75f0b28c900a1333d" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.631106 4848 scope.go:117] "RemoveContainer" containerID="272ebb608ee1f2bb5c58907aab4f4947a04ae59529614952c2994ff490528f1b" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.685655 4848 scope.go:117] "RemoveContainer" containerID="c940759c43b01c94d60d81a1069f82c0b54af5b13c2891edb642549a6774735a" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.766866 4848 scope.go:117] "RemoveContainer" containerID="ebc3b009cf446e78b02aa4807272e0de9c1e2ac6667948d1e45f6b12e02c568b" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.828904 4848 scope.go:117] "RemoveContainer" containerID="7ab4def029874b47a7238ffe5351fd78acdc1d4792b252fd28ff5214af48c1ca" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.877540 4848 scope.go:117] "RemoveContainer" containerID="e8eee4a623c4f7cb747ee23a1e57a2c62d6eb3b9701ec420784603cdcbac0d12" Dec 04 14:24:36 crc kubenswrapper[4848]: I1204 14:24:36.900266 4848 scope.go:117] "RemoveContainer" containerID="bc9161643dba2afa8240d7ac32bb8827495bd79e26bd160d7ba4c91469643dc2" Dec 04 14:24:38 crc kubenswrapper[4848]: I1204 14:24:38.034817 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-fb6e-account-create-update-sdrw9"] Dec 04 14:24:38 crc kubenswrapper[4848]: I1204 14:24:38.048701 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-fb6e-account-create-update-sdrw9"] Dec 04 14:24:38 crc kubenswrapper[4848]: I1204 14:24:38.407734 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6b33655-7b9c-4663-b8d0-85be8361f65f" path="/var/lib/kubelet/pods/d6b33655-7b9c-4663-b8d0-85be8361f65f/volumes" Dec 04 14:24:44 crc kubenswrapper[4848]: I1204 14:24:44.045086 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-s9gd9"] Dec 04 14:24:44 crc kubenswrapper[4848]: I1204 14:24:44.057085 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-s9gd9"] Dec 04 14:24:44 crc kubenswrapper[4848]: I1204 14:24:44.408607 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="618e60f3-6ba7-435e-b112-8ac2e7c6e5d9" path="/var/lib/kubelet/pods/618e60f3-6ba7-435e-b112-8ac2e7c6e5d9/volumes" Dec 04 14:24:47 crc kubenswrapper[4848]: I1204 14:24:47.036085 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8tc6k"] Dec 04 14:24:47 crc kubenswrapper[4848]: I1204 14:24:47.045541 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8tc6k"] Dec 04 14:24:48 crc kubenswrapper[4848]: I1204 14:24:48.407058 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb391cd3-6f78-4e51-ac06-14c9420d0f06" path="/var/lib/kubelet/pods/cb391cd3-6f78-4e51-ac06-14c9420d0f06/volumes" Dec 04 14:24:49 crc kubenswrapper[4848]: I1204 14:24:49.896327 4848 generic.go:334] "Generic (PLEG): container finished" podID="9d69a925-9c1f-4405-9300-e6c4c3cada23" containerID="aadc4b5b043c1c9343fd04fde8366c5c837d2ce9c2dce94b1c64be0e3edfd285" exitCode=0 Dec 04 14:24:49 crc kubenswrapper[4848]: I1204 14:24:49.896441 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" event={"ID":"9d69a925-9c1f-4405-9300-e6c4c3cada23","Type":"ContainerDied","Data":"aadc4b5b043c1c9343fd04fde8366c5c837d2ce9c2dce94b1c64be0e3edfd285"} Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.400647 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.483143 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-inventory\") pod \"9d69a925-9c1f-4405-9300-e6c4c3cada23\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.483241 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-ssh-key\") pod \"9d69a925-9c1f-4405-9300-e6c4c3cada23\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.483421 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5ngm\" (UniqueName: \"kubernetes.io/projected/9d69a925-9c1f-4405-9300-e6c4c3cada23-kube-api-access-s5ngm\") pod \"9d69a925-9c1f-4405-9300-e6c4c3cada23\" (UID: \"9d69a925-9c1f-4405-9300-e6c4c3cada23\") " Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.489475 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d69a925-9c1f-4405-9300-e6c4c3cada23-kube-api-access-s5ngm" (OuterVolumeSpecName: "kube-api-access-s5ngm") pod "9d69a925-9c1f-4405-9300-e6c4c3cada23" (UID: "9d69a925-9c1f-4405-9300-e6c4c3cada23"). InnerVolumeSpecName "kube-api-access-s5ngm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.521011 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9d69a925-9c1f-4405-9300-e6c4c3cada23" (UID: "9d69a925-9c1f-4405-9300-e6c4c3cada23"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.543402 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-inventory" (OuterVolumeSpecName: "inventory") pod "9d69a925-9c1f-4405-9300-e6c4c3cada23" (UID: "9d69a925-9c1f-4405-9300-e6c4c3cada23"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.585084 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.585118 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5ngm\" (UniqueName: \"kubernetes.io/projected/9d69a925-9c1f-4405-9300-e6c4c3cada23-kube-api-access-s5ngm\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.585133 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d69a925-9c1f-4405-9300-e6c4c3cada23-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.941764 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" event={"ID":"9d69a925-9c1f-4405-9300-e6c4c3cada23","Type":"ContainerDied","Data":"b39de2be5fe8ec92c1a4e9c3023715bf1128fe0ad1964d27789e0a28c7ab7ea4"} Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.941805 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b39de2be5fe8ec92c1a4e9c3023715bf1128fe0ad1964d27789e0a28c7ab7ea4" Dec 04 14:24:51 crc kubenswrapper[4848]: I1204 14:24:51.941814 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.017548 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9"] Dec 04 14:24:52 crc kubenswrapper[4848]: E1204 14:24:52.018069 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="extract-utilities" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.018092 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="extract-utilities" Dec 04 14:24:52 crc kubenswrapper[4848]: E1204 14:24:52.018109 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d69a925-9c1f-4405-9300-e6c4c3cada23" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.018119 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d69a925-9c1f-4405-9300-e6c4c3cada23" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 14:24:52 crc kubenswrapper[4848]: E1204 14:24:52.018152 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="registry-server" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.018160 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="registry-server" Dec 04 14:24:52 crc kubenswrapper[4848]: E1204 14:24:52.018196 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="extract-content" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.018203 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="extract-content" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.021936 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae352946-5e42-4d28-87d0-389cea7430ae" containerName="registry-server" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.022072 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d69a925-9c1f-4405-9300-e6c4c3cada23" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.023979 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.026431 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.026640 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.026805 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.027807 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.029918 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9"] Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.096701 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpkbx\" (UniqueName: \"kubernetes.io/projected/fbf57421-8b0a-42bd-a682-65afd34f39bb-kube-api-access-qpkbx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.097053 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.097133 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.201021 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.201132 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.201317 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpkbx\" (UniqueName: \"kubernetes.io/projected/fbf57421-8b0a-42bd-a682-65afd34f39bb-kube-api-access-qpkbx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.206011 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.206604 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.221751 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpkbx\" (UniqueName: \"kubernetes.io/projected/fbf57421-8b0a-42bd-a682-65afd34f39bb-kube-api-access-qpkbx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.358414 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:24:52 crc kubenswrapper[4848]: I1204 14:24:52.993386 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9"] Dec 04 14:24:53 crc kubenswrapper[4848]: I1204 14:24:53.962853 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" event={"ID":"fbf57421-8b0a-42bd-a682-65afd34f39bb","Type":"ContainerStarted","Data":"2a1a7458135522193e449e19f960358eb9604a01e8b0c9365f2f3a5273597416"} Dec 04 14:24:53 crc kubenswrapper[4848]: I1204 14:24:53.963726 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" event={"ID":"fbf57421-8b0a-42bd-a682-65afd34f39bb","Type":"ContainerStarted","Data":"0681eea36f7f2616abdcf1bee5e227c1d655c4238d7ce60816310cef8901c899"} Dec 04 14:24:53 crc kubenswrapper[4848]: I1204 14:24:53.988466 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" podStartSLOduration=2.560301796 podStartE2EDuration="2.988443227s" podCreationTimestamp="2025-12-04 14:24:51 +0000 UTC" firstStartedPulling="2025-12-04 14:24:52.994230309 +0000 UTC m=+2196.936726837" lastFinishedPulling="2025-12-04 14:24:53.42237174 +0000 UTC m=+2197.364868268" observedRunningTime="2025-12-04 14:24:53.986724244 +0000 UTC m=+2197.929220772" watchObservedRunningTime="2025-12-04 14:24:53.988443227 +0000 UTC m=+2197.930939755" Dec 04 14:25:29 crc kubenswrapper[4848]: I1204 14:25:29.062424 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-msdsj"] Dec 04 14:25:29 crc kubenswrapper[4848]: I1204 14:25:29.077700 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-msdsj"] Dec 04 14:25:30 crc kubenswrapper[4848]: I1204 14:25:30.407623 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1607ed31-5235-4f0c-a8e9-90f428f1da21" path="/var/lib/kubelet/pods/1607ed31-5235-4f0c-a8e9-90f428f1da21/volumes" Dec 04 14:25:37 crc kubenswrapper[4848]: I1204 14:25:37.069521 4848 scope.go:117] "RemoveContainer" containerID="1a0303464165f9c664f35f880812155bdf885db86f9276f9bb584f140fe87c8a" Dec 04 14:25:37 crc kubenswrapper[4848]: I1204 14:25:37.105088 4848 scope.go:117] "RemoveContainer" containerID="68bd1d83369d6855828689b87806360391bad180ef3c9372f79e96228be122f4" Dec 04 14:25:37 crc kubenswrapper[4848]: I1204 14:25:37.176873 4848 scope.go:117] "RemoveContainer" containerID="e6bb0c6f78150ddbeb78da3883863401a04086ec3f231a8cdcfeaf559006f080" Dec 04 14:25:37 crc kubenswrapper[4848]: I1204 14:25:37.225719 4848 scope.go:117] "RemoveContainer" containerID="efc73cb5821d728230b71deab4febbbdd8471088fbe6454ec39fa4a2f68e23ac" Dec 04 14:26:11 crc kubenswrapper[4848]: I1204 14:26:11.893785 4848 generic.go:334] "Generic (PLEG): container finished" podID="fbf57421-8b0a-42bd-a682-65afd34f39bb" containerID="2a1a7458135522193e449e19f960358eb9604a01e8b0c9365f2f3a5273597416" exitCode=0 Dec 04 14:26:11 crc kubenswrapper[4848]: I1204 14:26:11.893962 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" event={"ID":"fbf57421-8b0a-42bd-a682-65afd34f39bb","Type":"ContainerDied","Data":"2a1a7458135522193e449e19f960358eb9604a01e8b0c9365f2f3a5273597416"} Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.458706 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.660801 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-inventory\") pod \"fbf57421-8b0a-42bd-a682-65afd34f39bb\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.660877 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-ssh-key\") pod \"fbf57421-8b0a-42bd-a682-65afd34f39bb\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.661068 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpkbx\" (UniqueName: \"kubernetes.io/projected/fbf57421-8b0a-42bd-a682-65afd34f39bb-kube-api-access-qpkbx\") pod \"fbf57421-8b0a-42bd-a682-65afd34f39bb\" (UID: \"fbf57421-8b0a-42bd-a682-65afd34f39bb\") " Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.666309 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf57421-8b0a-42bd-a682-65afd34f39bb-kube-api-access-qpkbx" (OuterVolumeSpecName: "kube-api-access-qpkbx") pod "fbf57421-8b0a-42bd-a682-65afd34f39bb" (UID: "fbf57421-8b0a-42bd-a682-65afd34f39bb"). InnerVolumeSpecName "kube-api-access-qpkbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.694413 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-inventory" (OuterVolumeSpecName: "inventory") pod "fbf57421-8b0a-42bd-a682-65afd34f39bb" (UID: "fbf57421-8b0a-42bd-a682-65afd34f39bb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.696238 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fbf57421-8b0a-42bd-a682-65afd34f39bb" (UID: "fbf57421-8b0a-42bd-a682-65afd34f39bb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.766131 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpkbx\" (UniqueName: \"kubernetes.io/projected/fbf57421-8b0a-42bd-a682-65afd34f39bb-kube-api-access-qpkbx\") on node \"crc\" DevicePath \"\"" Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.766202 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.766214 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbf57421-8b0a-42bd-a682-65afd34f39bb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.919303 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" event={"ID":"fbf57421-8b0a-42bd-a682-65afd34f39bb","Type":"ContainerDied","Data":"0681eea36f7f2616abdcf1bee5e227c1d655c4238d7ce60816310cef8901c899"} Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.919608 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0681eea36f7f2616abdcf1bee5e227c1d655c4238d7ce60816310cef8901c899" Dec 04 14:26:13 crc kubenswrapper[4848]: I1204 14:26:13.919401 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.011339 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt"] Dec 04 14:26:14 crc kubenswrapper[4848]: E1204 14:26:14.011799 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf57421-8b0a-42bd-a682-65afd34f39bb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.011819 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf57421-8b0a-42bd-a682-65afd34f39bb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.012085 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf57421-8b0a-42bd-a682-65afd34f39bb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.012835 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.015226 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.015364 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.015367 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.015756 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.027999 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt"] Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.076431 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ggjl\" (UniqueName: \"kubernetes.io/projected/bae5190f-858d-4fdc-8116-78911f113311-kube-api-access-7ggjl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.076503 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.076587 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.182055 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.183382 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.184414 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ggjl\" (UniqueName: \"kubernetes.io/projected/bae5190f-858d-4fdc-8116-78911f113311-kube-api-access-7ggjl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.192350 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.192921 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.204556 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ggjl\" (UniqueName: \"kubernetes.io/projected/bae5190f-858d-4fdc-8116-78911f113311-kube-api-access-7ggjl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.314390 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.314767 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.332778 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.886760 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt"] Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.894463 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:26:14 crc kubenswrapper[4848]: I1204 14:26:14.929839 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" event={"ID":"bae5190f-858d-4fdc-8116-78911f113311","Type":"ContainerStarted","Data":"77eb1a052f4b996511a64d6687a0f16ac702eba60f4fca94d1f7d62b69c66610"} Dec 04 14:26:15 crc kubenswrapper[4848]: I1204 14:26:15.946805 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" event={"ID":"bae5190f-858d-4fdc-8116-78911f113311","Type":"ContainerStarted","Data":"54a6dbbc7940d3b7b0117535c19f25e9f682d4523218ca50c97702eedfc8ba73"} Dec 04 14:26:20 crc kubenswrapper[4848]: I1204 14:26:20.996545 4848 generic.go:334] "Generic (PLEG): container finished" podID="bae5190f-858d-4fdc-8116-78911f113311" containerID="54a6dbbc7940d3b7b0117535c19f25e9f682d4523218ca50c97702eedfc8ba73" exitCode=0 Dec 04 14:26:20 crc kubenswrapper[4848]: I1204 14:26:20.996705 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" event={"ID":"bae5190f-858d-4fdc-8116-78911f113311","Type":"ContainerDied","Data":"54a6dbbc7940d3b7b0117535c19f25e9f682d4523218ca50c97702eedfc8ba73"} Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.483308 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.684205 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-ssh-key\") pod \"bae5190f-858d-4fdc-8116-78911f113311\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.684318 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-inventory\") pod \"bae5190f-858d-4fdc-8116-78911f113311\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.684370 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ggjl\" (UniqueName: \"kubernetes.io/projected/bae5190f-858d-4fdc-8116-78911f113311-kube-api-access-7ggjl\") pod \"bae5190f-858d-4fdc-8116-78911f113311\" (UID: \"bae5190f-858d-4fdc-8116-78911f113311\") " Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.710236 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae5190f-858d-4fdc-8116-78911f113311-kube-api-access-7ggjl" (OuterVolumeSpecName: "kube-api-access-7ggjl") pod "bae5190f-858d-4fdc-8116-78911f113311" (UID: "bae5190f-858d-4fdc-8116-78911f113311"). InnerVolumeSpecName "kube-api-access-7ggjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.747170 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bae5190f-858d-4fdc-8116-78911f113311" (UID: "bae5190f-858d-4fdc-8116-78911f113311"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.759427 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-inventory" (OuterVolumeSpecName: "inventory") pod "bae5190f-858d-4fdc-8116-78911f113311" (UID: "bae5190f-858d-4fdc-8116-78911f113311"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.787336 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ggjl\" (UniqueName: \"kubernetes.io/projected/bae5190f-858d-4fdc-8116-78911f113311-kube-api-access-7ggjl\") on node \"crc\" DevicePath \"\"" Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.787370 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:26:22 crc kubenswrapper[4848]: I1204 14:26:22.787380 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bae5190f-858d-4fdc-8116-78911f113311-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.019600 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" event={"ID":"bae5190f-858d-4fdc-8116-78911f113311","Type":"ContainerDied","Data":"77eb1a052f4b996511a64d6687a0f16ac702eba60f4fca94d1f7d62b69c66610"} Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.019682 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77eb1a052f4b996511a64d6687a0f16ac702eba60f4fca94d1f7d62b69c66610" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.019702 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.099491 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr"] Dec 04 14:26:23 crc kubenswrapper[4848]: E1204 14:26:23.100201 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae5190f-858d-4fdc-8116-78911f113311" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.100234 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae5190f-858d-4fdc-8116-78911f113311" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.100590 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae5190f-858d-4fdc-8116-78911f113311" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.101665 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.105652 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.105745 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.105883 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.105998 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.112588 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr"] Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.195892 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n85ck\" (UniqueName: \"kubernetes.io/projected/51fab24f-2ea8-430c-810b-a538832876de-kube-api-access-n85ck\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.196023 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.196074 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.298198 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n85ck\" (UniqueName: \"kubernetes.io/projected/51fab24f-2ea8-430c-810b-a538832876de-kube-api-access-n85ck\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.298289 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.298339 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.303472 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.303634 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.313053 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n85ck\" (UniqueName: \"kubernetes.io/projected/51fab24f-2ea8-430c-810b-a538832876de-kube-api-access-n85ck\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gbtwr\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.437056 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:26:23 crc kubenswrapper[4848]: I1204 14:26:23.985672 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr"] Dec 04 14:26:24 crc kubenswrapper[4848]: I1204 14:26:24.032575 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" event={"ID":"51fab24f-2ea8-430c-810b-a538832876de","Type":"ContainerStarted","Data":"557db207d88aab63604cf6331e5301b4313ba44568cfef589867d58de2b99b3c"} Dec 04 14:26:25 crc kubenswrapper[4848]: I1204 14:26:25.044216 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" event={"ID":"51fab24f-2ea8-430c-810b-a538832876de","Type":"ContainerStarted","Data":"97d5df129aac41593b0ba17454cbb4c71a46a405db14225c57111103e599c73e"} Dec 04 14:26:25 crc kubenswrapper[4848]: I1204 14:26:25.069042 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" podStartSLOduration=1.648942936 podStartE2EDuration="2.069021942s" podCreationTimestamp="2025-12-04 14:26:23 +0000 UTC" firstStartedPulling="2025-12-04 14:26:23.99418241 +0000 UTC m=+2287.936678938" lastFinishedPulling="2025-12-04 14:26:24.414261416 +0000 UTC m=+2288.356757944" observedRunningTime="2025-12-04 14:26:25.059721055 +0000 UTC m=+2289.002217583" watchObservedRunningTime="2025-12-04 14:26:25.069021942 +0000 UTC m=+2289.011518470" Dec 04 14:26:44 crc kubenswrapper[4848]: I1204 14:26:44.314455 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:26:44 crc kubenswrapper[4848]: I1204 14:26:44.315034 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:27:06 crc kubenswrapper[4848]: I1204 14:27:06.501087 4848 generic.go:334] "Generic (PLEG): container finished" podID="51fab24f-2ea8-430c-810b-a538832876de" containerID="97d5df129aac41593b0ba17454cbb4c71a46a405db14225c57111103e599c73e" exitCode=0 Dec 04 14:27:06 crc kubenswrapper[4848]: I1204 14:27:06.501302 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" event={"ID":"51fab24f-2ea8-430c-810b-a538832876de","Type":"ContainerDied","Data":"97d5df129aac41593b0ba17454cbb4c71a46a405db14225c57111103e599c73e"} Dec 04 14:27:07 crc kubenswrapper[4848]: I1204 14:27:07.966151 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.077776 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-inventory\") pod \"51fab24f-2ea8-430c-810b-a538832876de\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.078114 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n85ck\" (UniqueName: \"kubernetes.io/projected/51fab24f-2ea8-430c-810b-a538832876de-kube-api-access-n85ck\") pod \"51fab24f-2ea8-430c-810b-a538832876de\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.078285 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-ssh-key\") pod \"51fab24f-2ea8-430c-810b-a538832876de\" (UID: \"51fab24f-2ea8-430c-810b-a538832876de\") " Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.085289 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51fab24f-2ea8-430c-810b-a538832876de-kube-api-access-n85ck" (OuterVolumeSpecName: "kube-api-access-n85ck") pod "51fab24f-2ea8-430c-810b-a538832876de" (UID: "51fab24f-2ea8-430c-810b-a538832876de"). InnerVolumeSpecName "kube-api-access-n85ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.109207 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-inventory" (OuterVolumeSpecName: "inventory") pod "51fab24f-2ea8-430c-810b-a538832876de" (UID: "51fab24f-2ea8-430c-810b-a538832876de"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.111829 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "51fab24f-2ea8-430c-810b-a538832876de" (UID: "51fab24f-2ea8-430c-810b-a538832876de"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.181356 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.181410 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n85ck\" (UniqueName: \"kubernetes.io/projected/51fab24f-2ea8-430c-810b-a538832876de-kube-api-access-n85ck\") on node \"crc\" DevicePath \"\"" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.181432 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51fab24f-2ea8-430c-810b-a538832876de-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.523393 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" event={"ID":"51fab24f-2ea8-430c-810b-a538832876de","Type":"ContainerDied","Data":"557db207d88aab63604cf6331e5301b4313ba44568cfef589867d58de2b99b3c"} Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.523440 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="557db207d88aab63604cf6331e5301b4313ba44568cfef589867d58de2b99b3c" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.523471 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gbtwr" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.622883 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt"] Dec 04 14:27:08 crc kubenswrapper[4848]: E1204 14:27:08.623460 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fab24f-2ea8-430c-810b-a538832876de" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.623482 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fab24f-2ea8-430c-810b-a538832876de" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.623793 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="51fab24f-2ea8-430c-810b-a538832876de" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.625065 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.628297 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.629320 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.629919 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.630769 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.639060 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt"] Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.693458 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.693868 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.694065 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xskwd\" (UniqueName: \"kubernetes.io/projected/8c603058-8c35-4c40-b700-2c358d0a5df7-kube-api-access-xskwd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.797095 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.797585 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xskwd\" (UniqueName: \"kubernetes.io/projected/8c603058-8c35-4c40-b700-2c358d0a5df7-kube-api-access-xskwd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.798021 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.807834 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.807907 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.817034 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xskwd\" (UniqueName: \"kubernetes.io/projected/8c603058-8c35-4c40-b700-2c358d0a5df7-kube-api-access-xskwd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:08 crc kubenswrapper[4848]: I1204 14:27:08.947011 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:27:09 crc kubenswrapper[4848]: I1204 14:27:09.537613 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt"] Dec 04 14:27:10 crc kubenswrapper[4848]: I1204 14:27:10.544232 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" event={"ID":"8c603058-8c35-4c40-b700-2c358d0a5df7","Type":"ContainerStarted","Data":"983089892b5e187bfaf08de20236e389f6c20827f26d98e93f5e0d3e45b9c395"} Dec 04 14:27:11 crc kubenswrapper[4848]: I1204 14:27:11.558873 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" event={"ID":"8c603058-8c35-4c40-b700-2c358d0a5df7","Type":"ContainerStarted","Data":"14e79444a420221bfe7dd3805b47e2b68b1b34a3b64c07f0a071d1c078b6568f"} Dec 04 14:27:11 crc kubenswrapper[4848]: I1204 14:27:11.585139 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" podStartSLOduration=2.666243689 podStartE2EDuration="3.585112908s" podCreationTimestamp="2025-12-04 14:27:08 +0000 UTC" firstStartedPulling="2025-12-04 14:27:09.548130526 +0000 UTC m=+2333.490627054" lastFinishedPulling="2025-12-04 14:27:10.466999745 +0000 UTC m=+2334.409496273" observedRunningTime="2025-12-04 14:27:11.583419157 +0000 UTC m=+2335.525915695" watchObservedRunningTime="2025-12-04 14:27:11.585112908 +0000 UTC m=+2335.527609446" Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.313879 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.314307 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.314381 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.315623 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.316015 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" gracePeriod=600 Dec 04 14:27:14 crc kubenswrapper[4848]: E1204 14:27:14.470214 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.598459 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" exitCode=0 Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.598537 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6"} Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.598600 4848 scope.go:117] "RemoveContainer" containerID="6068840f9ec731a3fcec181636bc276aedf6fad5ca2e8f03411d734389322f62" Dec 04 14:27:14 crc kubenswrapper[4848]: I1204 14:27:14.600380 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:27:14 crc kubenswrapper[4848]: E1204 14:27:14.601143 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:27:28 crc kubenswrapper[4848]: I1204 14:27:28.393441 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:27:28 crc kubenswrapper[4848]: E1204 14:27:28.394213 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:27:39 crc kubenswrapper[4848]: I1204 14:27:39.057739 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-9h9bt"] Dec 04 14:27:39 crc kubenswrapper[4848]: I1204 14:27:39.068775 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-9h9bt"] Dec 04 14:27:40 crc kubenswrapper[4848]: I1204 14:27:40.408211 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa" path="/var/lib/kubelet/pods/cb507b26-2c92-49e1-a3a4-ca0bc6d9ebaa/volumes" Dec 04 14:27:43 crc kubenswrapper[4848]: I1204 14:27:43.395302 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:27:43 crc kubenswrapper[4848]: E1204 14:27:43.396527 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:27:54 crc kubenswrapper[4848]: I1204 14:27:54.394118 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:27:54 crc kubenswrapper[4848]: E1204 14:27:54.394902 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:28:01 crc kubenswrapper[4848]: I1204 14:28:01.116868 4848 generic.go:334] "Generic (PLEG): container finished" podID="8c603058-8c35-4c40-b700-2c358d0a5df7" containerID="14e79444a420221bfe7dd3805b47e2b68b1b34a3b64c07f0a071d1c078b6568f" exitCode=0 Dec 04 14:28:01 crc kubenswrapper[4848]: I1204 14:28:01.117014 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" event={"ID":"8c603058-8c35-4c40-b700-2c358d0a5df7","Type":"ContainerDied","Data":"14e79444a420221bfe7dd3805b47e2b68b1b34a3b64c07f0a071d1c078b6568f"} Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.644270 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.811855 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-inventory\") pod \"8c603058-8c35-4c40-b700-2c358d0a5df7\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.812390 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-ssh-key\") pod \"8c603058-8c35-4c40-b700-2c358d0a5df7\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.812455 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xskwd\" (UniqueName: \"kubernetes.io/projected/8c603058-8c35-4c40-b700-2c358d0a5df7-kube-api-access-xskwd\") pod \"8c603058-8c35-4c40-b700-2c358d0a5df7\" (UID: \"8c603058-8c35-4c40-b700-2c358d0a5df7\") " Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.821051 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c603058-8c35-4c40-b700-2c358d0a5df7-kube-api-access-xskwd" (OuterVolumeSpecName: "kube-api-access-xskwd") pod "8c603058-8c35-4c40-b700-2c358d0a5df7" (UID: "8c603058-8c35-4c40-b700-2c358d0a5df7"). InnerVolumeSpecName "kube-api-access-xskwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.856386 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8c603058-8c35-4c40-b700-2c358d0a5df7" (UID: "8c603058-8c35-4c40-b700-2c358d0a5df7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.879875 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-inventory" (OuterVolumeSpecName: "inventory") pod "8c603058-8c35-4c40-b700-2c358d0a5df7" (UID: "8c603058-8c35-4c40-b700-2c358d0a5df7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.916404 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xskwd\" (UniqueName: \"kubernetes.io/projected/8c603058-8c35-4c40-b700-2c358d0a5df7-kube-api-access-xskwd\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.916459 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:02 crc kubenswrapper[4848]: I1204 14:28:02.916471 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c603058-8c35-4c40-b700-2c358d0a5df7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.143427 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" event={"ID":"8c603058-8c35-4c40-b700-2c358d0a5df7","Type":"ContainerDied","Data":"983089892b5e187bfaf08de20236e389f6c20827f26d98e93f5e0d3e45b9c395"} Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.143479 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="983089892b5e187bfaf08de20236e389f6c20827f26d98e93f5e0d3e45b9c395" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.143567 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.239395 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dxn92"] Dec 04 14:28:03 crc kubenswrapper[4848]: E1204 14:28:03.240010 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c603058-8c35-4c40-b700-2c358d0a5df7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.240029 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c603058-8c35-4c40-b700-2c358d0a5df7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.240289 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c603058-8c35-4c40-b700-2c358d0a5df7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.241216 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.252790 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dxn92"] Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.255393 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.257091 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.257243 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.257499 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.333506 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.333783 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.333932 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxtkd\" (UniqueName: \"kubernetes.io/projected/7ae92c99-2f33-44d5-a478-9a685afd0f89-kube-api-access-cxtkd\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.436037 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.436444 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.436654 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxtkd\" (UniqueName: \"kubernetes.io/projected/7ae92c99-2f33-44d5-a478-9a685afd0f89-kube-api-access-cxtkd\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.440351 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.442264 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.456230 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxtkd\" (UniqueName: \"kubernetes.io/projected/7ae92c99-2f33-44d5-a478-9a685afd0f89-kube-api-access-cxtkd\") pod \"ssh-known-hosts-edpm-deployment-dxn92\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:03 crc kubenswrapper[4848]: I1204 14:28:03.572894 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:04 crc kubenswrapper[4848]: I1204 14:28:04.134806 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dxn92"] Dec 04 14:28:04 crc kubenswrapper[4848]: I1204 14:28:04.161282 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" event={"ID":"7ae92c99-2f33-44d5-a478-9a685afd0f89","Type":"ContainerStarted","Data":"0e33803ece8b45aaf6953b7e4aabcf701d0b69e708ee31160a7952c460c64999"} Dec 04 14:28:06 crc kubenswrapper[4848]: I1204 14:28:06.185107 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" event={"ID":"7ae92c99-2f33-44d5-a478-9a685afd0f89","Type":"ContainerStarted","Data":"9ec857a7c8c5237e32eafc6f5f880abfde9bffce40c1922b542d824709c2c60e"} Dec 04 14:28:06 crc kubenswrapper[4848]: I1204 14:28:06.206897 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" podStartSLOduration=2.730348018 podStartE2EDuration="3.206879658s" podCreationTimestamp="2025-12-04 14:28:03 +0000 UTC" firstStartedPulling="2025-12-04 14:28:04.14005338 +0000 UTC m=+2388.082549898" lastFinishedPulling="2025-12-04 14:28:04.616585 +0000 UTC m=+2388.559081538" observedRunningTime="2025-12-04 14:28:06.200764959 +0000 UTC m=+2390.143261497" watchObservedRunningTime="2025-12-04 14:28:06.206879658 +0000 UTC m=+2390.149376186" Dec 04 14:28:07 crc kubenswrapper[4848]: I1204 14:28:07.395103 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:28:07 crc kubenswrapper[4848]: E1204 14:28:07.395543 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:28:12 crc kubenswrapper[4848]: I1204 14:28:12.253417 4848 generic.go:334] "Generic (PLEG): container finished" podID="7ae92c99-2f33-44d5-a478-9a685afd0f89" containerID="9ec857a7c8c5237e32eafc6f5f880abfde9bffce40c1922b542d824709c2c60e" exitCode=0 Dec 04 14:28:12 crc kubenswrapper[4848]: I1204 14:28:12.253521 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" event={"ID":"7ae92c99-2f33-44d5-a478-9a685afd0f89","Type":"ContainerDied","Data":"9ec857a7c8c5237e32eafc6f5f880abfde9bffce40c1922b542d824709c2c60e"} Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.708375 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.787455 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxtkd\" (UniqueName: \"kubernetes.io/projected/7ae92c99-2f33-44d5-a478-9a685afd0f89-kube-api-access-cxtkd\") pod \"7ae92c99-2f33-44d5-a478-9a685afd0f89\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.787561 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-ssh-key-openstack-edpm-ipam\") pod \"7ae92c99-2f33-44d5-a478-9a685afd0f89\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.787663 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-inventory-0\") pod \"7ae92c99-2f33-44d5-a478-9a685afd0f89\" (UID: \"7ae92c99-2f33-44d5-a478-9a685afd0f89\") " Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.793603 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ae92c99-2f33-44d5-a478-9a685afd0f89-kube-api-access-cxtkd" (OuterVolumeSpecName: "kube-api-access-cxtkd") pod "7ae92c99-2f33-44d5-a478-9a685afd0f89" (UID: "7ae92c99-2f33-44d5-a478-9a685afd0f89"). InnerVolumeSpecName "kube-api-access-cxtkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.819431 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7ae92c99-2f33-44d5-a478-9a685afd0f89" (UID: "7ae92c99-2f33-44d5-a478-9a685afd0f89"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.822546 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "7ae92c99-2f33-44d5-a478-9a685afd0f89" (UID: "7ae92c99-2f33-44d5-a478-9a685afd0f89"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.891364 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxtkd\" (UniqueName: \"kubernetes.io/projected/7ae92c99-2f33-44d5-a478-9a685afd0f89-kube-api-access-cxtkd\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.891678 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:13 crc kubenswrapper[4848]: I1204 14:28:13.891755 4848 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7ae92c99-2f33-44d5-a478-9a685afd0f89-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.275938 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" event={"ID":"7ae92c99-2f33-44d5-a478-9a685afd0f89","Type":"ContainerDied","Data":"0e33803ece8b45aaf6953b7e4aabcf701d0b69e708ee31160a7952c460c64999"} Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.275990 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e33803ece8b45aaf6953b7e4aabcf701d0b69e708ee31160a7952c460c64999" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.276033 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dxn92" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.362160 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc"] Dec 04 14:28:14 crc kubenswrapper[4848]: E1204 14:28:14.362738 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae92c99-2f33-44d5-a478-9a685afd0f89" containerName="ssh-known-hosts-edpm-deployment" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.362761 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae92c99-2f33-44d5-a478-9a685afd0f89" containerName="ssh-known-hosts-edpm-deployment" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.363158 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ae92c99-2f33-44d5-a478-9a685afd0f89" containerName="ssh-known-hosts-edpm-deployment" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.364231 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.366822 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.366928 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.367572 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.369366 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.378537 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc"] Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.505572 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.505656 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q6sk\" (UniqueName: \"kubernetes.io/projected/5b8a5833-af2e-4fef-9bfa-bbcec7064538-kube-api-access-5q6sk\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.505921 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.608360 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.608465 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q6sk\" (UniqueName: \"kubernetes.io/projected/5b8a5833-af2e-4fef-9bfa-bbcec7064538-kube-api-access-5q6sk\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.608567 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.612221 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.613072 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.626433 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q6sk\" (UniqueName: \"kubernetes.io/projected/5b8a5833-af2e-4fef-9bfa-bbcec7064538-kube-api-access-5q6sk\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dfmnc\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:14 crc kubenswrapper[4848]: I1204 14:28:14.720542 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:15 crc kubenswrapper[4848]: I1204 14:28:15.264627 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc"] Dec 04 14:28:15 crc kubenswrapper[4848]: I1204 14:28:15.287243 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" event={"ID":"5b8a5833-af2e-4fef-9bfa-bbcec7064538","Type":"ContainerStarted","Data":"a3c934585f2ca912989152d7478b1b092d83d24820b6de32d869d5dc38691561"} Dec 04 14:28:16 crc kubenswrapper[4848]: I1204 14:28:16.299693 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" event={"ID":"5b8a5833-af2e-4fef-9bfa-bbcec7064538","Type":"ContainerStarted","Data":"9e93f24a8a5eaff7f833f127052d78e1da795e7ebee95560f24b802a4aeb5bb6"} Dec 04 14:28:16 crc kubenswrapper[4848]: I1204 14:28:16.321137 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" podStartSLOduration=1.893349666 podStartE2EDuration="2.321118231s" podCreationTimestamp="2025-12-04 14:28:14 +0000 UTC" firstStartedPulling="2025-12-04 14:28:15.268990971 +0000 UTC m=+2399.211487499" lastFinishedPulling="2025-12-04 14:28:15.696759506 +0000 UTC m=+2399.639256064" observedRunningTime="2025-12-04 14:28:16.315855773 +0000 UTC m=+2400.258352311" watchObservedRunningTime="2025-12-04 14:28:16.321118231 +0000 UTC m=+2400.263614759" Dec 04 14:28:21 crc kubenswrapper[4848]: I1204 14:28:21.394590 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:28:21 crc kubenswrapper[4848]: E1204 14:28:21.395993 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:28:24 crc kubenswrapper[4848]: I1204 14:28:24.384019 4848 generic.go:334] "Generic (PLEG): container finished" podID="5b8a5833-af2e-4fef-9bfa-bbcec7064538" containerID="9e93f24a8a5eaff7f833f127052d78e1da795e7ebee95560f24b802a4aeb5bb6" exitCode=0 Dec 04 14:28:24 crc kubenswrapper[4848]: I1204 14:28:24.384106 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" event={"ID":"5b8a5833-af2e-4fef-9bfa-bbcec7064538","Type":"ContainerDied","Data":"9e93f24a8a5eaff7f833f127052d78e1da795e7ebee95560f24b802a4aeb5bb6"} Dec 04 14:28:25 crc kubenswrapper[4848]: I1204 14:28:25.844934 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.051614 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-ssh-key\") pod \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.051900 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-inventory\") pod \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.052074 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q6sk\" (UniqueName: \"kubernetes.io/projected/5b8a5833-af2e-4fef-9bfa-bbcec7064538-kube-api-access-5q6sk\") pod \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\" (UID: \"5b8a5833-af2e-4fef-9bfa-bbcec7064538\") " Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.065380 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b8a5833-af2e-4fef-9bfa-bbcec7064538-kube-api-access-5q6sk" (OuterVolumeSpecName: "kube-api-access-5q6sk") pod "5b8a5833-af2e-4fef-9bfa-bbcec7064538" (UID: "5b8a5833-af2e-4fef-9bfa-bbcec7064538"). InnerVolumeSpecName "kube-api-access-5q6sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.091715 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-inventory" (OuterVolumeSpecName: "inventory") pod "5b8a5833-af2e-4fef-9bfa-bbcec7064538" (UID: "5b8a5833-af2e-4fef-9bfa-bbcec7064538"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.093405 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5b8a5833-af2e-4fef-9bfa-bbcec7064538" (UID: "5b8a5833-af2e-4fef-9bfa-bbcec7064538"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.155195 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q6sk\" (UniqueName: \"kubernetes.io/projected/5b8a5833-af2e-4fef-9bfa-bbcec7064538-kube-api-access-5q6sk\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.155254 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.155275 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5b8a5833-af2e-4fef-9bfa-bbcec7064538-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.411097 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" event={"ID":"5b8a5833-af2e-4fef-9bfa-bbcec7064538","Type":"ContainerDied","Data":"a3c934585f2ca912989152d7478b1b092d83d24820b6de32d869d5dc38691561"} Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.411145 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3c934585f2ca912989152d7478b1b092d83d24820b6de32d869d5dc38691561" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.411152 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dfmnc" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.492402 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf"] Dec 04 14:28:26 crc kubenswrapper[4848]: E1204 14:28:26.492889 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8a5833-af2e-4fef-9bfa-bbcec7064538" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.492908 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8a5833-af2e-4fef-9bfa-bbcec7064538" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.493164 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b8a5833-af2e-4fef-9bfa-bbcec7064538" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.494041 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.497603 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.497786 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.497937 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.503520 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.505132 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf"] Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.563710 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55vx4\" (UniqueName: \"kubernetes.io/projected/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-kube-api-access-55vx4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.564199 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.564339 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.666262 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.666339 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.666407 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55vx4\" (UniqueName: \"kubernetes.io/projected/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-kube-api-access-55vx4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.671911 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.674893 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.686389 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55vx4\" (UniqueName: \"kubernetes.io/projected/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-kube-api-access-55vx4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:26 crc kubenswrapper[4848]: I1204 14:28:26.818177 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:27 crc kubenswrapper[4848]: I1204 14:28:27.062014 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-6jlbh"] Dec 04 14:28:27 crc kubenswrapper[4848]: I1204 14:28:27.071491 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-6jlbh"] Dec 04 14:28:27 crc kubenswrapper[4848]: I1204 14:28:27.411072 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf"] Dec 04 14:28:28 crc kubenswrapper[4848]: I1204 14:28:28.407612 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf1c92f4-2158-4669-9816-cfda5923549a" path="/var/lib/kubelet/pods/cf1c92f4-2158-4669-9816-cfda5923549a/volumes" Dec 04 14:28:28 crc kubenswrapper[4848]: I1204 14:28:28.433179 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" event={"ID":"2af1d16b-4257-4fcd-b6fa-dc37b1313e40","Type":"ContainerStarted","Data":"4828046412d77ac3b5d8637e754b7d91b38835e99f6149d968d0426d5bfa3e0b"} Dec 04 14:28:29 crc kubenswrapper[4848]: I1204 14:28:29.446062 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" event={"ID":"2af1d16b-4257-4fcd-b6fa-dc37b1313e40","Type":"ContainerStarted","Data":"0f1a490cdb2f992b9bc8a4763a8b44eb6d0ad334eb28348d857100d4607c867a"} Dec 04 14:28:29 crc kubenswrapper[4848]: I1204 14:28:29.466701 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" podStartSLOduration=2.26450494 podStartE2EDuration="3.466682699s" podCreationTimestamp="2025-12-04 14:28:26 +0000 UTC" firstStartedPulling="2025-12-04 14:28:27.416883965 +0000 UTC m=+2411.359380493" lastFinishedPulling="2025-12-04 14:28:28.619061724 +0000 UTC m=+2412.561558252" observedRunningTime="2025-12-04 14:28:29.466355271 +0000 UTC m=+2413.408851889" watchObservedRunningTime="2025-12-04 14:28:29.466682699 +0000 UTC m=+2413.409179227" Dec 04 14:28:32 crc kubenswrapper[4848]: I1204 14:28:32.393721 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:28:32 crc kubenswrapper[4848]: E1204 14:28:32.394342 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:28:37 crc kubenswrapper[4848]: I1204 14:28:37.405088 4848 scope.go:117] "RemoveContainer" containerID="e8e46ba09e90621e97561c2ff1278ce17460794bbe14aef860e25f72e8596e6c" Dec 04 14:28:37 crc kubenswrapper[4848]: I1204 14:28:37.455686 4848 scope.go:117] "RemoveContainer" containerID="967008c19d4c8753fdcb441e40778498f8829fd91e80e8babe4b842c6d36816e" Dec 04 14:28:39 crc kubenswrapper[4848]: I1204 14:28:39.566574 4848 generic.go:334] "Generic (PLEG): container finished" podID="2af1d16b-4257-4fcd-b6fa-dc37b1313e40" containerID="0f1a490cdb2f992b9bc8a4763a8b44eb6d0ad334eb28348d857100d4607c867a" exitCode=0 Dec 04 14:28:39 crc kubenswrapper[4848]: I1204 14:28:39.566675 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" event={"ID":"2af1d16b-4257-4fcd-b6fa-dc37b1313e40","Type":"ContainerDied","Data":"0f1a490cdb2f992b9bc8a4763a8b44eb6d0ad334eb28348d857100d4607c867a"} Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.067236 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.220336 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55vx4\" (UniqueName: \"kubernetes.io/projected/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-kube-api-access-55vx4\") pod \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.220677 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-ssh-key\") pod \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.220729 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-inventory\") pod \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\" (UID: \"2af1d16b-4257-4fcd-b6fa-dc37b1313e40\") " Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.225688 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-kube-api-access-55vx4" (OuterVolumeSpecName: "kube-api-access-55vx4") pod "2af1d16b-4257-4fcd-b6fa-dc37b1313e40" (UID: "2af1d16b-4257-4fcd-b6fa-dc37b1313e40"). InnerVolumeSpecName "kube-api-access-55vx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.252822 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2af1d16b-4257-4fcd-b6fa-dc37b1313e40" (UID: "2af1d16b-4257-4fcd-b6fa-dc37b1313e40"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.255797 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-inventory" (OuterVolumeSpecName: "inventory") pod "2af1d16b-4257-4fcd-b6fa-dc37b1313e40" (UID: "2af1d16b-4257-4fcd-b6fa-dc37b1313e40"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.324049 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.324096 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.324110 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55vx4\" (UniqueName: \"kubernetes.io/projected/2af1d16b-4257-4fcd-b6fa-dc37b1313e40-kube-api-access-55vx4\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.586506 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" event={"ID":"2af1d16b-4257-4fcd-b6fa-dc37b1313e40","Type":"ContainerDied","Data":"4828046412d77ac3b5d8637e754b7d91b38835e99f6149d968d0426d5bfa3e0b"} Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.586542 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4828046412d77ac3b5d8637e754b7d91b38835e99f6149d968d0426d5bfa3e0b" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.586542 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.690086 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp"] Dec 04 14:28:41 crc kubenswrapper[4848]: E1204 14:28:41.690901 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af1d16b-4257-4fcd-b6fa-dc37b1313e40" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.690921 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af1d16b-4257-4fcd-b6fa-dc37b1313e40" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.691200 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af1d16b-4257-4fcd-b6fa-dc37b1313e40" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.692064 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.694934 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.695133 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.695214 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.699472 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.699917 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.699920 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.700358 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.700143 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.700196 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.709137 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp"] Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.734537 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.734903 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.735081 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.735213 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.735304 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.735393 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.735572 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.735737 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.735856 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.735971 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.736070 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.736176 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw52v\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-kube-api-access-gw52v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.736324 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.736442 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.736553 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.736677 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838102 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838183 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838220 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838238 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838258 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838289 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw52v\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-kube-api-access-gw52v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838328 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838346 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838372 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838408 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838451 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838473 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838503 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838544 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838563 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.838581 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.842399 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.843153 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.843529 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.843819 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.845897 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.845916 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.846362 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.846805 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.847090 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.848342 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.849940 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.850484 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.850715 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.853547 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.854882 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:41 crc kubenswrapper[4848]: I1204 14:28:41.857543 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw52v\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-kube-api-access-gw52v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:42 crc kubenswrapper[4848]: I1204 14:28:42.011906 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:28:42 crc kubenswrapper[4848]: I1204 14:28:42.519869 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp"] Dec 04 14:28:42 crc kubenswrapper[4848]: I1204 14:28:42.598463 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" event={"ID":"fe59edb1-8c86-40a3-b68a-7f597af859de","Type":"ContainerStarted","Data":"b3d9594bb50cd9674354385a98890e9bea13af460551d2a10ede366bc9b4b68b"} Dec 04 14:28:43 crc kubenswrapper[4848]: I1204 14:28:43.610667 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" event={"ID":"fe59edb1-8c86-40a3-b68a-7f597af859de","Type":"ContainerStarted","Data":"5d31d401719af2598bbf8793cf5cef08d3c54ea7baa617279c3fc102e4fb3771"} Dec 04 14:28:43 crc kubenswrapper[4848]: I1204 14:28:43.638048 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" podStartSLOduration=2.128778792 podStartE2EDuration="2.638026137s" podCreationTimestamp="2025-12-04 14:28:41 +0000 UTC" firstStartedPulling="2025-12-04 14:28:42.525304184 +0000 UTC m=+2426.467800712" lastFinishedPulling="2025-12-04 14:28:43.034551529 +0000 UTC m=+2426.977048057" observedRunningTime="2025-12-04 14:28:43.630479793 +0000 UTC m=+2427.572976341" watchObservedRunningTime="2025-12-04 14:28:43.638026137 +0000 UTC m=+2427.580522665" Dec 04 14:28:47 crc kubenswrapper[4848]: I1204 14:28:47.394244 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:28:47 crc kubenswrapper[4848]: E1204 14:28:47.395171 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:29:00 crc kubenswrapper[4848]: I1204 14:29:00.394238 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:29:00 crc kubenswrapper[4848]: E1204 14:29:00.395319 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:29:11 crc kubenswrapper[4848]: I1204 14:29:11.393964 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:29:11 crc kubenswrapper[4848]: E1204 14:29:11.394888 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.068094 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rt4g9"] Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.071291 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.079842 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rt4g9"] Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.240635 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tjc4\" (UniqueName: \"kubernetes.io/projected/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-kube-api-access-6tjc4\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.241075 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-catalog-content\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.241535 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-utilities\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.344149 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-utilities\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.344272 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tjc4\" (UniqueName: \"kubernetes.io/projected/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-kube-api-access-6tjc4\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.344373 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-catalog-content\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.345506 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-utilities\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.345550 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-catalog-content\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.374358 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tjc4\" (UniqueName: \"kubernetes.io/projected/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-kube-api-access-6tjc4\") pod \"certified-operators-rt4g9\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.393611 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:21 crc kubenswrapper[4848]: I1204 14:29:21.965813 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rt4g9"] Dec 04 14:29:22 crc kubenswrapper[4848]: I1204 14:29:22.026307 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rt4g9" event={"ID":"4c53fb4a-ebe3-4503-9353-21da2b0e16c8","Type":"ContainerStarted","Data":"073284775b06fd33206b0ab475c61aeaa325e0f862ac53cd7a230cf59f89ddd2"} Dec 04 14:29:23 crc kubenswrapper[4848]: I1204 14:29:23.039327 4848 generic.go:334] "Generic (PLEG): container finished" podID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerID="b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c" exitCode=0 Dec 04 14:29:23 crc kubenswrapper[4848]: I1204 14:29:23.039444 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rt4g9" event={"ID":"4c53fb4a-ebe3-4503-9353-21da2b0e16c8","Type":"ContainerDied","Data":"b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c"} Dec 04 14:29:25 crc kubenswrapper[4848]: I1204 14:29:25.060153 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rt4g9" event={"ID":"4c53fb4a-ebe3-4503-9353-21da2b0e16c8","Type":"ContainerStarted","Data":"a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9"} Dec 04 14:29:25 crc kubenswrapper[4848]: I1204 14:29:25.393543 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:29:25 crc kubenswrapper[4848]: E1204 14:29:25.394254 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:29:26 crc kubenswrapper[4848]: I1204 14:29:26.072081 4848 generic.go:334] "Generic (PLEG): container finished" podID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerID="a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9" exitCode=0 Dec 04 14:29:26 crc kubenswrapper[4848]: I1204 14:29:26.072136 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rt4g9" event={"ID":"4c53fb4a-ebe3-4503-9353-21da2b0e16c8","Type":"ContainerDied","Data":"a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9"} Dec 04 14:29:27 crc kubenswrapper[4848]: I1204 14:29:27.084204 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rt4g9" event={"ID":"4c53fb4a-ebe3-4503-9353-21da2b0e16c8","Type":"ContainerStarted","Data":"dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a"} Dec 04 14:29:27 crc kubenswrapper[4848]: I1204 14:29:27.114865 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rt4g9" podStartSLOduration=2.639470788 podStartE2EDuration="6.114840924s" podCreationTimestamp="2025-12-04 14:29:21 +0000 UTC" firstStartedPulling="2025-12-04 14:29:23.041776281 +0000 UTC m=+2466.984272809" lastFinishedPulling="2025-12-04 14:29:26.517146417 +0000 UTC m=+2470.459642945" observedRunningTime="2025-12-04 14:29:27.104159925 +0000 UTC m=+2471.046656453" watchObservedRunningTime="2025-12-04 14:29:27.114840924 +0000 UTC m=+2471.057337452" Dec 04 14:29:31 crc kubenswrapper[4848]: I1204 14:29:31.121140 4848 generic.go:334] "Generic (PLEG): container finished" podID="fe59edb1-8c86-40a3-b68a-7f597af859de" containerID="5d31d401719af2598bbf8793cf5cef08d3c54ea7baa617279c3fc102e4fb3771" exitCode=0 Dec 04 14:29:31 crc kubenswrapper[4848]: I1204 14:29:31.121223 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" event={"ID":"fe59edb1-8c86-40a3-b68a-7f597af859de","Type":"ContainerDied","Data":"5d31d401719af2598bbf8793cf5cef08d3c54ea7baa617279c3fc102e4fb3771"} Dec 04 14:29:31 crc kubenswrapper[4848]: I1204 14:29:31.394022 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:31 crc kubenswrapper[4848]: I1204 14:29:31.394331 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:31 crc kubenswrapper[4848]: I1204 14:29:31.438885 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.189987 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.620007 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.727939 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-neutron-metadata-combined-ca-bundle\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728095 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728124 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-libvirt-combined-ca-bundle\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728188 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728217 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-nova-combined-ca-bundle\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728272 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw52v\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-kube-api-access-gw52v\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728334 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-repo-setup-combined-ca-bundle\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728365 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ovn-combined-ca-bundle\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728388 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-bootstrap-combined-ca-bundle\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728447 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ssh-key\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728589 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728804 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-ovn-default-certs-0\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728827 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-inventory\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728876 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-power-monitoring-combined-ca-bundle\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728900 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.728929 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-combined-ca-bundle\") pod \"fe59edb1-8c86-40a3-b68a-7f597af859de\" (UID: \"fe59edb1-8c86-40a3-b68a-7f597af859de\") " Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.737556 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.737606 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-kube-api-access-gw52v" (OuterVolumeSpecName: "kube-api-access-gw52v") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "kube-api-access-gw52v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.738046 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.738046 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.738580 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.738617 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.750225 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.750285 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.750307 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.750324 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.750334 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.750661 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.750691 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.752290 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.775528 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.789689 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-inventory" (OuterVolumeSpecName: "inventory") pod "fe59edb1-8c86-40a3-b68a-7f597af859de" (UID: "fe59edb1-8c86-40a3-b68a-7f597af859de"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832402 4848 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832466 4848 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832481 4848 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832493 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832506 4848 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832521 4848 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832535 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832547 4848 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832557 4848 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832566 4848 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832577 4848 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832588 4848 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832599 4848 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832610 4848 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832620 4848 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe59edb1-8c86-40a3-b68a-7f597af859de-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.832631 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw52v\" (UniqueName: \"kubernetes.io/projected/fe59edb1-8c86-40a3-b68a-7f597af859de-kube-api-access-gw52v\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:32 crc kubenswrapper[4848]: I1204 14:29:32.852249 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rt4g9"] Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.144728 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" event={"ID":"fe59edb1-8c86-40a3-b68a-7f597af859de","Type":"ContainerDied","Data":"b3d9594bb50cd9674354385a98890e9bea13af460551d2a10ede366bc9b4b68b"} Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.144776 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3d9594bb50cd9674354385a98890e9bea13af460551d2a10ede366bc9b4b68b" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.144823 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.276929 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb"] Dec 04 14:29:33 crc kubenswrapper[4848]: E1204 14:29:33.277909 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe59edb1-8c86-40a3-b68a-7f597af859de" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.277969 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe59edb1-8c86-40a3-b68a-7f597af859de" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.278311 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe59edb1-8c86-40a3-b68a-7f597af859de" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.279630 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.281773 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.281832 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.282235 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.286027 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.286067 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.291571 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb"] Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.343733 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.343803 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.343830 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.343945 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.344423 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzkrz\" (UniqueName: \"kubernetes.io/projected/54ce5df0-f586-463e-ad53-6ddc68e8ce90-kube-api-access-xzkrz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.447147 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzkrz\" (UniqueName: \"kubernetes.io/projected/54ce5df0-f586-463e-ad53-6ddc68e8ce90-kube-api-access-xzkrz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.447359 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.447404 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.447445 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.447494 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.448580 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.453596 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.454586 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.463578 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.472835 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzkrz\" (UniqueName: \"kubernetes.io/projected/54ce5df0-f586-463e-ad53-6ddc68e8ce90-kube-api-access-xzkrz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l9mmb\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:33 crc kubenswrapper[4848]: I1204 14:29:33.609586 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:29:34 crc kubenswrapper[4848]: I1204 14:29:34.161234 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rt4g9" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerName="registry-server" containerID="cri-o://dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a" gracePeriod=2 Dec 04 14:29:34 crc kubenswrapper[4848]: I1204 14:29:34.185620 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb"] Dec 04 14:29:34 crc kubenswrapper[4848]: E1204 14:29:34.464797 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c53fb4a_ebe3_4503_9353_21da2b0e16c8.slice/crio-dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a.scope\": RecentStats: unable to find data in memory cache]" Dec 04 14:29:34 crc kubenswrapper[4848]: I1204 14:29:34.853992 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:34 crc kubenswrapper[4848]: I1204 14:29:34.980401 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-catalog-content\") pod \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " Dec 04 14:29:34 crc kubenswrapper[4848]: I1204 14:29:34.980568 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-utilities\") pod \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " Dec 04 14:29:34 crc kubenswrapper[4848]: I1204 14:29:34.981165 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tjc4\" (UniqueName: \"kubernetes.io/projected/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-kube-api-access-6tjc4\") pod \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\" (UID: \"4c53fb4a-ebe3-4503-9353-21da2b0e16c8\") " Dec 04 14:29:34 crc kubenswrapper[4848]: I1204 14:29:34.982222 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-utilities" (OuterVolumeSpecName: "utilities") pod "4c53fb4a-ebe3-4503-9353-21da2b0e16c8" (UID: "4c53fb4a-ebe3-4503-9353-21da2b0e16c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:29:34 crc kubenswrapper[4848]: I1204 14:29:34.985592 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-kube-api-access-6tjc4" (OuterVolumeSpecName: "kube-api-access-6tjc4") pod "4c53fb4a-ebe3-4503-9353-21da2b0e16c8" (UID: "4c53fb4a-ebe3-4503-9353-21da2b0e16c8"). InnerVolumeSpecName "kube-api-access-6tjc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.033798 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c53fb4a-ebe3-4503-9353-21da2b0e16c8" (UID: "4c53fb4a-ebe3-4503-9353-21da2b0e16c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.083839 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tjc4\" (UniqueName: \"kubernetes.io/projected/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-kube-api-access-6tjc4\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.083873 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.083882 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c53fb4a-ebe3-4503-9353-21da2b0e16c8-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.175307 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" event={"ID":"54ce5df0-f586-463e-ad53-6ddc68e8ce90","Type":"ContainerStarted","Data":"ae27aa76626eb230e42f59fde279d2adc3439ae21bb6e963901a790c12b3a811"} Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.175361 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" event={"ID":"54ce5df0-f586-463e-ad53-6ddc68e8ce90","Type":"ContainerStarted","Data":"39a83b3e54fd00ac5cbd99a6b05d265259fd5aa16740db214f76a6d6cf4383f0"} Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.178877 4848 generic.go:334] "Generic (PLEG): container finished" podID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerID="dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a" exitCode=0 Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.178934 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rt4g9" event={"ID":"4c53fb4a-ebe3-4503-9353-21da2b0e16c8","Type":"ContainerDied","Data":"dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a"} Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.178982 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rt4g9" event={"ID":"4c53fb4a-ebe3-4503-9353-21da2b0e16c8","Type":"ContainerDied","Data":"073284775b06fd33206b0ab475c61aeaa325e0f862ac53cd7a230cf59f89ddd2"} Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.179007 4848 scope.go:117] "RemoveContainer" containerID="dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.179136 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rt4g9" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.201264 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" podStartSLOduration=1.729598497 podStartE2EDuration="2.201244568s" podCreationTimestamp="2025-12-04 14:29:33 +0000 UTC" firstStartedPulling="2025-12-04 14:29:34.193010676 +0000 UTC m=+2478.135507204" lastFinishedPulling="2025-12-04 14:29:34.664656747 +0000 UTC m=+2478.607153275" observedRunningTime="2025-12-04 14:29:35.193471198 +0000 UTC m=+2479.135967736" watchObservedRunningTime="2025-12-04 14:29:35.201244568 +0000 UTC m=+2479.143741096" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.211684 4848 scope.go:117] "RemoveContainer" containerID="a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.223797 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rt4g9"] Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.233810 4848 scope.go:117] "RemoveContainer" containerID="b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.235412 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rt4g9"] Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.258415 4848 scope.go:117] "RemoveContainer" containerID="dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a" Dec 04 14:29:35 crc kubenswrapper[4848]: E1204 14:29:35.259295 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a\": container with ID starting with dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a not found: ID does not exist" containerID="dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.259330 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a"} err="failed to get container status \"dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a\": rpc error: code = NotFound desc = could not find container \"dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a\": container with ID starting with dd118a16de63cd8f07956f79394ede29726e1caf6113c1d905f9cd8b19db2e8a not found: ID does not exist" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.259351 4848 scope.go:117] "RemoveContainer" containerID="a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9" Dec 04 14:29:35 crc kubenswrapper[4848]: E1204 14:29:35.259785 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9\": container with ID starting with a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9 not found: ID does not exist" containerID="a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.259812 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9"} err="failed to get container status \"a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9\": rpc error: code = NotFound desc = could not find container \"a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9\": container with ID starting with a4dcc1d1ed7754f3999f5077eac9056c7cfaec3b74b5c3d1c23b7c4fe56804c9 not found: ID does not exist" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.259829 4848 scope.go:117] "RemoveContainer" containerID="b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c" Dec 04 14:29:35 crc kubenswrapper[4848]: E1204 14:29:35.260308 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c\": container with ID starting with b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c not found: ID does not exist" containerID="b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c" Dec 04 14:29:35 crc kubenswrapper[4848]: I1204 14:29:35.260332 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c"} err="failed to get container status \"b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c\": rpc error: code = NotFound desc = could not find container \"b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c\": container with ID starting with b9750a1448b3de2210ec42957032d10104697525dcca7dcb5df10b6d37044a4c not found: ID does not exist" Dec 04 14:29:36 crc kubenswrapper[4848]: I1204 14:29:36.409106 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" path="/var/lib/kubelet/pods/4c53fb4a-ebe3-4503-9353-21da2b0e16c8/volumes" Dec 04 14:29:40 crc kubenswrapper[4848]: I1204 14:29:40.394029 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:29:40 crc kubenswrapper[4848]: E1204 14:29:40.394854 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:29:51 crc kubenswrapper[4848]: I1204 14:29:51.393266 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:29:51 crc kubenswrapper[4848]: E1204 14:29:51.394099 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.326000 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm"] Dec 04 14:30:00 crc kubenswrapper[4848]: E1204 14:30:00.327646 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerName="extract-utilities" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.327666 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerName="extract-utilities" Dec 04 14:30:00 crc kubenswrapper[4848]: E1204 14:30:00.327676 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerName="extract-content" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.327683 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerName="extract-content" Dec 04 14:30:00 crc kubenswrapper[4848]: E1204 14:30:00.327705 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerName="registry-server" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.327712 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerName="registry-server" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.328043 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c53fb4a-ebe3-4503-9353-21da2b0e16c8" containerName="registry-server" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.333209 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.335666 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.336650 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm"] Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.336801 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.484212 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-config-volume\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.484264 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-secret-volume\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.484353 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f8qb\" (UniqueName: \"kubernetes.io/projected/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-kube-api-access-8f8qb\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.587306 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-config-volume\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.587587 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-secret-volume\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.587743 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f8qb\" (UniqueName: \"kubernetes.io/projected/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-kube-api-access-8f8qb\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.588678 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-config-volume\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.602838 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-secret-volume\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.605848 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f8qb\" (UniqueName: \"kubernetes.io/projected/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-kube-api-access-8f8qb\") pod \"collect-profiles-29414310-259bm\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:00 crc kubenswrapper[4848]: I1204 14:30:00.668137 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:01 crc kubenswrapper[4848]: I1204 14:30:01.175159 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm"] Dec 04 14:30:01 crc kubenswrapper[4848]: I1204 14:30:01.453069 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" event={"ID":"be9c3abb-a5b2-4e89-8c9d-04b475f528ed","Type":"ContainerStarted","Data":"ed6f2e078870290acd286ff601629f0742643a3c53bafc05fd47d689b81eb818"} Dec 04 14:30:01 crc kubenswrapper[4848]: I1204 14:30:01.454068 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" event={"ID":"be9c3abb-a5b2-4e89-8c9d-04b475f528ed","Type":"ContainerStarted","Data":"bba0134ab2ae0731ccffed76025b64434a1674124607c21ff1a77d01a89955cb"} Dec 04 14:30:01 crc kubenswrapper[4848]: I1204 14:30:01.477596 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" podStartSLOduration=1.477569135 podStartE2EDuration="1.477569135s" podCreationTimestamp="2025-12-04 14:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:30:01.46666639 +0000 UTC m=+2505.409162918" watchObservedRunningTime="2025-12-04 14:30:01.477569135 +0000 UTC m=+2505.420065673" Dec 04 14:30:02 crc kubenswrapper[4848]: I1204 14:30:02.470200 4848 generic.go:334] "Generic (PLEG): container finished" podID="be9c3abb-a5b2-4e89-8c9d-04b475f528ed" containerID="ed6f2e078870290acd286ff601629f0742643a3c53bafc05fd47d689b81eb818" exitCode=0 Dec 04 14:30:02 crc kubenswrapper[4848]: I1204 14:30:02.470312 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" event={"ID":"be9c3abb-a5b2-4e89-8c9d-04b475f528ed","Type":"ContainerDied","Data":"ed6f2e078870290acd286ff601629f0742643a3c53bafc05fd47d689b81eb818"} Dec 04 14:30:03 crc kubenswrapper[4848]: I1204 14:30:03.864180 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:03 crc kubenswrapper[4848]: I1204 14:30:03.987679 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-config-volume\") pod \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " Dec 04 14:30:03 crc kubenswrapper[4848]: I1204 14:30:03.987880 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-secret-volume\") pod \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " Dec 04 14:30:03 crc kubenswrapper[4848]: I1204 14:30:03.987990 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f8qb\" (UniqueName: \"kubernetes.io/projected/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-kube-api-access-8f8qb\") pod \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\" (UID: \"be9c3abb-a5b2-4e89-8c9d-04b475f528ed\") " Dec 04 14:30:03 crc kubenswrapper[4848]: I1204 14:30:03.988627 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-config-volume" (OuterVolumeSpecName: "config-volume") pod "be9c3abb-a5b2-4e89-8c9d-04b475f528ed" (UID: "be9c3abb-a5b2-4e89-8c9d-04b475f528ed"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:30:03 crc kubenswrapper[4848]: I1204 14:30:03.994775 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "be9c3abb-a5b2-4e89-8c9d-04b475f528ed" (UID: "be9c3abb-a5b2-4e89-8c9d-04b475f528ed"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:03 crc kubenswrapper[4848]: I1204 14:30:03.994908 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-kube-api-access-8f8qb" (OuterVolumeSpecName: "kube-api-access-8f8qb") pod "be9c3abb-a5b2-4e89-8c9d-04b475f528ed" (UID: "be9c3abb-a5b2-4e89-8c9d-04b475f528ed"). InnerVolumeSpecName "kube-api-access-8f8qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:30:04 crc kubenswrapper[4848]: I1204 14:30:04.092072 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:04 crc kubenswrapper[4848]: I1204 14:30:04.092106 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:04 crc kubenswrapper[4848]: I1204 14:30:04.092137 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f8qb\" (UniqueName: \"kubernetes.io/projected/be9c3abb-a5b2-4e89-8c9d-04b475f528ed-kube-api-access-8f8qb\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:04 crc kubenswrapper[4848]: I1204 14:30:04.520188 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" event={"ID":"be9c3abb-a5b2-4e89-8c9d-04b475f528ed","Type":"ContainerDied","Data":"bba0134ab2ae0731ccffed76025b64434a1674124607c21ff1a77d01a89955cb"} Dec 04 14:30:04 crc kubenswrapper[4848]: I1204 14:30:04.520435 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bba0134ab2ae0731ccffed76025b64434a1674124607c21ff1a77d01a89955cb" Dec 04 14:30:04 crc kubenswrapper[4848]: I1204 14:30:04.520308 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm" Dec 04 14:30:04 crc kubenswrapper[4848]: I1204 14:30:04.580017 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm"] Dec 04 14:30:04 crc kubenswrapper[4848]: I1204 14:30:04.590937 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-d9rhm"] Dec 04 14:30:06 crc kubenswrapper[4848]: I1204 14:30:06.403074 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:30:06 crc kubenswrapper[4848]: E1204 14:30:06.403739 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:30:06 crc kubenswrapper[4848]: I1204 14:30:06.409680 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca4d77ac-6c91-4789-8a01-553fc2563d65" path="/var/lib/kubelet/pods/ca4d77ac-6c91-4789-8a01-553fc2563d65/volumes" Dec 04 14:30:18 crc kubenswrapper[4848]: I1204 14:30:18.394617 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:30:18 crc kubenswrapper[4848]: E1204 14:30:18.395428 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:30:31 crc kubenswrapper[4848]: I1204 14:30:31.393817 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:30:31 crc kubenswrapper[4848]: E1204 14:30:31.394454 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:30:37 crc kubenswrapper[4848]: I1204 14:30:37.633700 4848 scope.go:117] "RemoveContainer" containerID="4285e30e6e9757f3f2a3627eda467f9dda8d82ac97aca1918864f3a326c56878" Dec 04 14:30:38 crc kubenswrapper[4848]: I1204 14:30:38.864503 4848 generic.go:334] "Generic (PLEG): container finished" podID="54ce5df0-f586-463e-ad53-6ddc68e8ce90" containerID="ae27aa76626eb230e42f59fde279d2adc3439ae21bb6e963901a790c12b3a811" exitCode=0 Dec 04 14:30:38 crc kubenswrapper[4848]: I1204 14:30:38.864622 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" event={"ID":"54ce5df0-f586-463e-ad53-6ddc68e8ce90","Type":"ContainerDied","Data":"ae27aa76626eb230e42f59fde279d2adc3439ae21bb6e963901a790c12b3a811"} Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.325537 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.447907 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzkrz\" (UniqueName: \"kubernetes.io/projected/54ce5df0-f586-463e-ad53-6ddc68e8ce90-kube-api-access-xzkrz\") pod \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.448782 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovn-combined-ca-bundle\") pod \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.448985 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-inventory\") pod \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.449076 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovncontroller-config-0\") pod \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.449142 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ssh-key\") pod \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\" (UID: \"54ce5df0-f586-463e-ad53-6ddc68e8ce90\") " Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.454346 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ce5df0-f586-463e-ad53-6ddc68e8ce90-kube-api-access-xzkrz" (OuterVolumeSpecName: "kube-api-access-xzkrz") pod "54ce5df0-f586-463e-ad53-6ddc68e8ce90" (UID: "54ce5df0-f586-463e-ad53-6ddc68e8ce90"). InnerVolumeSpecName "kube-api-access-xzkrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.471699 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "54ce5df0-f586-463e-ad53-6ddc68e8ce90" (UID: "54ce5df0-f586-463e-ad53-6ddc68e8ce90"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.485407 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-inventory" (OuterVolumeSpecName: "inventory") pod "54ce5df0-f586-463e-ad53-6ddc68e8ce90" (UID: "54ce5df0-f586-463e-ad53-6ddc68e8ce90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.499577 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "54ce5df0-f586-463e-ad53-6ddc68e8ce90" (UID: "54ce5df0-f586-463e-ad53-6ddc68e8ce90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.502172 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "54ce5df0-f586-463e-ad53-6ddc68e8ce90" (UID: "54ce5df0-f586-463e-ad53-6ddc68e8ce90"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.553585 4848 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.553884 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.554084 4848 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.554217 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54ce5df0-f586-463e-ad53-6ddc68e8ce90-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.554342 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzkrz\" (UniqueName: \"kubernetes.io/projected/54ce5df0-f586-463e-ad53-6ddc68e8ce90-kube-api-access-xzkrz\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.887862 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" event={"ID":"54ce5df0-f586-463e-ad53-6ddc68e8ce90","Type":"ContainerDied","Data":"39a83b3e54fd00ac5cbd99a6b05d265259fd5aa16740db214f76a6d6cf4383f0"} Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.887943 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39a83b3e54fd00ac5cbd99a6b05d265259fd5aa16740db214f76a6d6cf4383f0" Dec 04 14:30:40 crc kubenswrapper[4848]: I1204 14:30:40.888117 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l9mmb" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.006038 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr"] Dec 04 14:30:41 crc kubenswrapper[4848]: E1204 14:30:41.006512 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9c3abb-a5b2-4e89-8c9d-04b475f528ed" containerName="collect-profiles" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.006524 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9c3abb-a5b2-4e89-8c9d-04b475f528ed" containerName="collect-profiles" Dec 04 14:30:41 crc kubenswrapper[4848]: E1204 14:30:41.006569 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ce5df0-f586-463e-ad53-6ddc68e8ce90" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.006576 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ce5df0-f586-463e-ad53-6ddc68e8ce90" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.006786 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="be9c3abb-a5b2-4e89-8c9d-04b475f528ed" containerName="collect-profiles" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.006796 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ce5df0-f586-463e-ad53-6ddc68e8ce90" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.007585 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.010444 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.012042 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.012088 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.012049 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.012218 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.012331 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.037131 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr"] Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.167384 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.167755 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.167833 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.168030 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.168409 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvpjr\" (UniqueName: \"kubernetes.io/projected/b1231a87-3a32-4b52-a5b0-d546f15523e1-kube-api-access-gvpjr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.168459 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.270699 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.270776 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.270924 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvpjr\" (UniqueName: \"kubernetes.io/projected/b1231a87-3a32-4b52-a5b0-d546f15523e1-kube-api-access-gvpjr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.270973 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.271049 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.271122 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.276421 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.276495 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.276809 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.277499 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.278127 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.289189 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvpjr\" (UniqueName: \"kubernetes.io/projected/b1231a87-3a32-4b52-a5b0-d546f15523e1-kube-api-access-gvpjr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.337189 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.880321 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr"] Dec 04 14:30:41 crc kubenswrapper[4848]: W1204 14:30:41.893173 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1231a87_3a32_4b52_a5b0_d546f15523e1.slice/crio-989fbb2749c471052abffe1a29392eacf4affe6312e9d4238f2682afc1efe01b WatchSource:0}: Error finding container 989fbb2749c471052abffe1a29392eacf4affe6312e9d4238f2682afc1efe01b: Status 404 returned error can't find the container with id 989fbb2749c471052abffe1a29392eacf4affe6312e9d4238f2682afc1efe01b Dec 04 14:30:41 crc kubenswrapper[4848]: I1204 14:30:41.903613 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" event={"ID":"b1231a87-3a32-4b52-a5b0-d546f15523e1","Type":"ContainerStarted","Data":"989fbb2749c471052abffe1a29392eacf4affe6312e9d4238f2682afc1efe01b"} Dec 04 14:30:42 crc kubenswrapper[4848]: I1204 14:30:42.915681 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" event={"ID":"b1231a87-3a32-4b52-a5b0-d546f15523e1","Type":"ContainerStarted","Data":"5657033902eb8598cdfb178743e2f48a6d7f7fa29da5ede690daae4335985005"} Dec 04 14:30:42 crc kubenswrapper[4848]: I1204 14:30:42.937809 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" podStartSLOduration=2.486649085 podStartE2EDuration="2.937790542s" podCreationTimestamp="2025-12-04 14:30:40 +0000 UTC" firstStartedPulling="2025-12-04 14:30:41.894671561 +0000 UTC m=+2545.837168089" lastFinishedPulling="2025-12-04 14:30:42.345813018 +0000 UTC m=+2546.288309546" observedRunningTime="2025-12-04 14:30:42.929755616 +0000 UTC m=+2546.872252144" watchObservedRunningTime="2025-12-04 14:30:42.937790542 +0000 UTC m=+2546.880287070" Dec 04 14:30:44 crc kubenswrapper[4848]: I1204 14:30:44.393369 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:30:44 crc kubenswrapper[4848]: E1204 14:30:44.393961 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:30:57 crc kubenswrapper[4848]: I1204 14:30:57.394157 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:30:57 crc kubenswrapper[4848]: E1204 14:30:57.395515 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.691781 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pq7kt"] Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.694817 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.711847 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pq7kt"] Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.879039 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-utilities\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.880107 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8s9c\" (UniqueName: \"kubernetes.io/projected/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-kube-api-access-x8s9c\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.880204 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-catalog-content\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.982146 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8s9c\" (UniqueName: \"kubernetes.io/projected/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-kube-api-access-x8s9c\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.982506 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-catalog-content\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.982696 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-utilities\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.983169 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-catalog-content\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:04 crc kubenswrapper[4848]: I1204 14:31:04.983207 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-utilities\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:05 crc kubenswrapper[4848]: I1204 14:31:04.999457 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8s9c\" (UniqueName: \"kubernetes.io/projected/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-kube-api-access-x8s9c\") pod \"community-operators-pq7kt\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:05 crc kubenswrapper[4848]: I1204 14:31:05.044348 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:05 crc kubenswrapper[4848]: W1204 14:31:05.619435 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ed05a4a_677e_43a6_a037_9bbb735bf5e6.slice/crio-eed43796312e6980144960cea09c17e8fb83dcfb288ebce8ede7fc9da2fb6d68 WatchSource:0}: Error finding container eed43796312e6980144960cea09c17e8fb83dcfb288ebce8ede7fc9da2fb6d68: Status 404 returned error can't find the container with id eed43796312e6980144960cea09c17e8fb83dcfb288ebce8ede7fc9da2fb6d68 Dec 04 14:31:05 crc kubenswrapper[4848]: I1204 14:31:05.619640 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pq7kt"] Dec 04 14:31:06 crc kubenswrapper[4848]: I1204 14:31:06.197792 4848 generic.go:334] "Generic (PLEG): container finished" podID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerID="e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a" exitCode=0 Dec 04 14:31:06 crc kubenswrapper[4848]: I1204 14:31:06.197896 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pq7kt" event={"ID":"8ed05a4a-677e-43a6-a037-9bbb735bf5e6","Type":"ContainerDied","Data":"e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a"} Dec 04 14:31:06 crc kubenswrapper[4848]: I1204 14:31:06.199234 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pq7kt" event={"ID":"8ed05a4a-677e-43a6-a037-9bbb735bf5e6","Type":"ContainerStarted","Data":"eed43796312e6980144960cea09c17e8fb83dcfb288ebce8ede7fc9da2fb6d68"} Dec 04 14:31:07 crc kubenswrapper[4848]: I1204 14:31:07.210608 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pq7kt" event={"ID":"8ed05a4a-677e-43a6-a037-9bbb735bf5e6","Type":"ContainerStarted","Data":"4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe"} Dec 04 14:31:08 crc kubenswrapper[4848]: I1204 14:31:08.395808 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:31:08 crc kubenswrapper[4848]: E1204 14:31:08.396713 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:31:09 crc kubenswrapper[4848]: I1204 14:31:09.238374 4848 generic.go:334] "Generic (PLEG): container finished" podID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerID="4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe" exitCode=0 Dec 04 14:31:09 crc kubenswrapper[4848]: I1204 14:31:09.238420 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pq7kt" event={"ID":"8ed05a4a-677e-43a6-a037-9bbb735bf5e6","Type":"ContainerDied","Data":"4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe"} Dec 04 14:31:11 crc kubenswrapper[4848]: I1204 14:31:11.261319 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pq7kt" event={"ID":"8ed05a4a-677e-43a6-a037-9bbb735bf5e6","Type":"ContainerStarted","Data":"fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a"} Dec 04 14:31:11 crc kubenswrapper[4848]: I1204 14:31:11.295281 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pq7kt" podStartSLOduration=3.372822839 podStartE2EDuration="7.295259812s" podCreationTimestamp="2025-12-04 14:31:04 +0000 UTC" firstStartedPulling="2025-12-04 14:31:06.200061633 +0000 UTC m=+2570.142558171" lastFinishedPulling="2025-12-04 14:31:10.122498616 +0000 UTC m=+2574.064995144" observedRunningTime="2025-12-04 14:31:11.28322476 +0000 UTC m=+2575.225721288" watchObservedRunningTime="2025-12-04 14:31:11.295259812 +0000 UTC m=+2575.237756340" Dec 04 14:31:15 crc kubenswrapper[4848]: I1204 14:31:15.044828 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:15 crc kubenswrapper[4848]: I1204 14:31:15.045488 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:15 crc kubenswrapper[4848]: I1204 14:31:15.119995 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:15 crc kubenswrapper[4848]: I1204 14:31:15.372418 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:15 crc kubenswrapper[4848]: I1204 14:31:15.417116 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pq7kt"] Dec 04 14:31:17 crc kubenswrapper[4848]: I1204 14:31:17.339092 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pq7kt" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerName="registry-server" containerID="cri-o://fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a" gracePeriod=2 Dec 04 14:31:17 crc kubenswrapper[4848]: I1204 14:31:17.884233 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.042304 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-utilities\") pod \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.042632 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-catalog-content\") pod \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.042665 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8s9c\" (UniqueName: \"kubernetes.io/projected/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-kube-api-access-x8s9c\") pod \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\" (UID: \"8ed05a4a-677e-43a6-a037-9bbb735bf5e6\") " Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.043292 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-utilities" (OuterVolumeSpecName: "utilities") pod "8ed05a4a-677e-43a6-a037-9bbb735bf5e6" (UID: "8ed05a4a-677e-43a6-a037-9bbb735bf5e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.043914 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.048534 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-kube-api-access-x8s9c" (OuterVolumeSpecName: "kube-api-access-x8s9c") pod "8ed05a4a-677e-43a6-a037-9bbb735bf5e6" (UID: "8ed05a4a-677e-43a6-a037-9bbb735bf5e6"). InnerVolumeSpecName "kube-api-access-x8s9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.102991 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ed05a4a-677e-43a6-a037-9bbb735bf5e6" (UID: "8ed05a4a-677e-43a6-a037-9bbb735bf5e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.146277 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.146560 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8s9c\" (UniqueName: \"kubernetes.io/projected/8ed05a4a-677e-43a6-a037-9bbb735bf5e6-kube-api-access-x8s9c\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.354189 4848 generic.go:334] "Generic (PLEG): container finished" podID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerID="fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a" exitCode=0 Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.354235 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pq7kt" event={"ID":"8ed05a4a-677e-43a6-a037-9bbb735bf5e6","Type":"ContainerDied","Data":"fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a"} Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.354252 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pq7kt" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.354274 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pq7kt" event={"ID":"8ed05a4a-677e-43a6-a037-9bbb735bf5e6","Type":"ContainerDied","Data":"eed43796312e6980144960cea09c17e8fb83dcfb288ebce8ede7fc9da2fb6d68"} Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.354293 4848 scope.go:117] "RemoveContainer" containerID="fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.377286 4848 scope.go:117] "RemoveContainer" containerID="4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.408098 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pq7kt"] Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.409960 4848 scope.go:117] "RemoveContainer" containerID="e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.413901 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pq7kt"] Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.458634 4848 scope.go:117] "RemoveContainer" containerID="fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a" Dec 04 14:31:18 crc kubenswrapper[4848]: E1204 14:31:18.459345 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a\": container with ID starting with fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a not found: ID does not exist" containerID="fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.459379 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a"} err="failed to get container status \"fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a\": rpc error: code = NotFound desc = could not find container \"fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a\": container with ID starting with fd628a0c28972400b01d5531b77f1a553a9e8022db5c42af84432b745827964a not found: ID does not exist" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.459403 4848 scope.go:117] "RemoveContainer" containerID="4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe" Dec 04 14:31:18 crc kubenswrapper[4848]: E1204 14:31:18.459845 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe\": container with ID starting with 4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe not found: ID does not exist" containerID="4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.459901 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe"} err="failed to get container status \"4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe\": rpc error: code = NotFound desc = could not find container \"4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe\": container with ID starting with 4c55d123278c2b7c09e5275274205c16cb4c148a4b8bd9675f9bd1b5afc3efbe not found: ID does not exist" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.459934 4848 scope.go:117] "RemoveContainer" containerID="e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a" Dec 04 14:31:18 crc kubenswrapper[4848]: E1204 14:31:18.460244 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a\": container with ID starting with e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a not found: ID does not exist" containerID="e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a" Dec 04 14:31:18 crc kubenswrapper[4848]: I1204 14:31:18.460278 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a"} err="failed to get container status \"e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a\": rpc error: code = NotFound desc = could not find container \"e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a\": container with ID starting with e3a59594662f47329bc245fc4f1e309c4c5cb7d54884aaae3beb35fec9e2554a not found: ID does not exist" Dec 04 14:31:19 crc kubenswrapper[4848]: I1204 14:31:19.394699 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:31:19 crc kubenswrapper[4848]: E1204 14:31:19.395316 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:31:20 crc kubenswrapper[4848]: I1204 14:31:20.406296 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" path="/var/lib/kubelet/pods/8ed05a4a-677e-43a6-a037-9bbb735bf5e6/volumes" Dec 04 14:31:30 crc kubenswrapper[4848]: I1204 14:31:30.394544 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:31:30 crc kubenswrapper[4848]: E1204 14:31:30.395357 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:31:30 crc kubenswrapper[4848]: I1204 14:31:30.500708 4848 generic.go:334] "Generic (PLEG): container finished" podID="b1231a87-3a32-4b52-a5b0-d546f15523e1" containerID="5657033902eb8598cdfb178743e2f48a6d7f7fa29da5ede690daae4335985005" exitCode=0 Dec 04 14:31:30 crc kubenswrapper[4848]: I1204 14:31:30.500755 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" event={"ID":"b1231a87-3a32-4b52-a5b0-d546f15523e1","Type":"ContainerDied","Data":"5657033902eb8598cdfb178743e2f48a6d7f7fa29da5ede690daae4335985005"} Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.004193 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.108679 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-ssh-key\") pod \"b1231a87-3a32-4b52-a5b0-d546f15523e1\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.108723 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-metadata-combined-ca-bundle\") pod \"b1231a87-3a32-4b52-a5b0-d546f15523e1\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.108759 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"b1231a87-3a32-4b52-a5b0-d546f15523e1\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.108878 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-nova-metadata-neutron-config-0\") pod \"b1231a87-3a32-4b52-a5b0-d546f15523e1\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.108961 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-inventory\") pod \"b1231a87-3a32-4b52-a5b0-d546f15523e1\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.109007 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvpjr\" (UniqueName: \"kubernetes.io/projected/b1231a87-3a32-4b52-a5b0-d546f15523e1-kube-api-access-gvpjr\") pod \"b1231a87-3a32-4b52-a5b0-d546f15523e1\" (UID: \"b1231a87-3a32-4b52-a5b0-d546f15523e1\") " Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.116094 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b1231a87-3a32-4b52-a5b0-d546f15523e1" (UID: "b1231a87-3a32-4b52-a5b0-d546f15523e1"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.116273 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1231a87-3a32-4b52-a5b0-d546f15523e1-kube-api-access-gvpjr" (OuterVolumeSpecName: "kube-api-access-gvpjr") pod "b1231a87-3a32-4b52-a5b0-d546f15523e1" (UID: "b1231a87-3a32-4b52-a5b0-d546f15523e1"). InnerVolumeSpecName "kube-api-access-gvpjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.141572 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "b1231a87-3a32-4b52-a5b0-d546f15523e1" (UID: "b1231a87-3a32-4b52-a5b0-d546f15523e1"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.141588 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-inventory" (OuterVolumeSpecName: "inventory") pod "b1231a87-3a32-4b52-a5b0-d546f15523e1" (UID: "b1231a87-3a32-4b52-a5b0-d546f15523e1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.143214 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "b1231a87-3a32-4b52-a5b0-d546f15523e1" (UID: "b1231a87-3a32-4b52-a5b0-d546f15523e1"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.150467 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b1231a87-3a32-4b52-a5b0-d546f15523e1" (UID: "b1231a87-3a32-4b52-a5b0-d546f15523e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.212188 4848 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.212236 4848 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.212251 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.212263 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvpjr\" (UniqueName: \"kubernetes.io/projected/b1231a87-3a32-4b52-a5b0-d546f15523e1-kube-api-access-gvpjr\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.212278 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.212290 4848 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1231a87-3a32-4b52-a5b0-d546f15523e1-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.531035 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" event={"ID":"b1231a87-3a32-4b52-a5b0-d546f15523e1","Type":"ContainerDied","Data":"989fbb2749c471052abffe1a29392eacf4affe6312e9d4238f2682afc1efe01b"} Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.531595 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="989fbb2749c471052abffe1a29392eacf4affe6312e9d4238f2682afc1efe01b" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.531153 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.605719 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2"] Dec 04 14:31:32 crc kubenswrapper[4848]: E1204 14:31:32.606228 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1231a87-3a32-4b52-a5b0-d546f15523e1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.606249 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1231a87-3a32-4b52-a5b0-d546f15523e1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 14:31:32 crc kubenswrapper[4848]: E1204 14:31:32.606268 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerName="extract-utilities" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.606276 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerName="extract-utilities" Dec 04 14:31:32 crc kubenswrapper[4848]: E1204 14:31:32.606310 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerName="extract-content" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.606316 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerName="extract-content" Dec 04 14:31:32 crc kubenswrapper[4848]: E1204 14:31:32.606336 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerName="registry-server" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.606343 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerName="registry-server" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.606558 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ed05a4a-677e-43a6-a037-9bbb735bf5e6" containerName="registry-server" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.606584 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1231a87-3a32-4b52-a5b0-d546f15523e1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.607383 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.609446 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.609741 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.610861 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.611399 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.611591 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.619342 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2"] Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.721355 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p489f\" (UniqueName: \"kubernetes.io/projected/3146cd9a-c65d-4942-bb4d-ee0951403d87-kube-api-access-p489f\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.721453 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.721547 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.721645 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.721717 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.824353 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.824548 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p489f\" (UniqueName: \"kubernetes.io/projected/3146cd9a-c65d-4942-bb4d-ee0951403d87-kube-api-access-p489f\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.824644 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.824724 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.824792 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.829549 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.830013 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.830695 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.831128 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.843183 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p489f\" (UniqueName: \"kubernetes.io/projected/3146cd9a-c65d-4942-bb4d-ee0951403d87-kube-api-access-p489f\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6lph2\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:32 crc kubenswrapper[4848]: I1204 14:31:32.933737 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:31:33 crc kubenswrapper[4848]: I1204 14:31:33.477868 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2"] Dec 04 14:31:33 crc kubenswrapper[4848]: I1204 14:31:33.489806 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:31:33 crc kubenswrapper[4848]: I1204 14:31:33.542823 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" event={"ID":"3146cd9a-c65d-4942-bb4d-ee0951403d87","Type":"ContainerStarted","Data":"e8feb2ca652ff2d7806fb7892c2a0b3498822424a1308da9d355bbb2ee2933f9"} Dec 04 14:31:34 crc kubenswrapper[4848]: I1204 14:31:34.556619 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" event={"ID":"3146cd9a-c65d-4942-bb4d-ee0951403d87","Type":"ContainerStarted","Data":"067d3258f855f684d35aa34fc141f010e30a9cd19fa44ff7c97397c6e4d9a44e"} Dec 04 14:31:34 crc kubenswrapper[4848]: I1204 14:31:34.574504 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" podStartSLOduration=2.080140227 podStartE2EDuration="2.574484275s" podCreationTimestamp="2025-12-04 14:31:32 +0000 UTC" firstStartedPulling="2025-12-04 14:31:33.48929717 +0000 UTC m=+2597.431793698" lastFinishedPulling="2025-12-04 14:31:33.983641178 +0000 UTC m=+2597.926137746" observedRunningTime="2025-12-04 14:31:34.569331489 +0000 UTC m=+2598.511828017" watchObservedRunningTime="2025-12-04 14:31:34.574484275 +0000 UTC m=+2598.516980803" Dec 04 14:31:43 crc kubenswrapper[4848]: I1204 14:31:43.393454 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:31:43 crc kubenswrapper[4848]: E1204 14:31:43.394176 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:31:58 crc kubenswrapper[4848]: I1204 14:31:58.393463 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:31:58 crc kubenswrapper[4848]: E1204 14:31:58.394479 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:32:13 crc kubenswrapper[4848]: I1204 14:32:13.394121 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:32:13 crc kubenswrapper[4848]: E1204 14:32:13.394986 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:32:25 crc kubenswrapper[4848]: I1204 14:32:25.393870 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:32:26 crc kubenswrapper[4848]: I1204 14:32:26.128858 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"ef540d9e8a947299436fb6249a1ab4f19679231bf019f3be88d55d65c9bd9249"} Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.063687 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mjjxj"] Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.068192 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.086238 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjjxj"] Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.174506 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-catalog-content\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.174603 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-utilities\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.174788 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brz45\" (UniqueName: \"kubernetes.io/projected/26bd9a14-0381-4383-a034-756b5be0e111-kube-api-access-brz45\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.277005 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-catalog-content\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.277079 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-utilities\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.277263 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brz45\" (UniqueName: \"kubernetes.io/projected/26bd9a14-0381-4383-a034-756b5be0e111-kube-api-access-brz45\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.277620 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-catalog-content\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.277639 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-utilities\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.304974 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brz45\" (UniqueName: \"kubernetes.io/projected/26bd9a14-0381-4383-a034-756b5be0e111-kube-api-access-brz45\") pod \"redhat-marketplace-mjjxj\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.394834 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:06 crc kubenswrapper[4848]: I1204 14:33:06.955126 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjjxj"] Dec 04 14:33:07 crc kubenswrapper[4848]: I1204 14:33:07.625508 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjjxj" event={"ID":"26bd9a14-0381-4383-a034-756b5be0e111","Type":"ContainerStarted","Data":"086c76789e6e84fa612057ec1047e311a3501558a99cf0fa6462ddcc3302046c"} Dec 04 14:33:08 crc kubenswrapper[4848]: I1204 14:33:08.636274 4848 generic.go:334] "Generic (PLEG): container finished" podID="26bd9a14-0381-4383-a034-756b5be0e111" containerID="fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1" exitCode=0 Dec 04 14:33:08 crc kubenswrapper[4848]: I1204 14:33:08.636325 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjjxj" event={"ID":"26bd9a14-0381-4383-a034-756b5be0e111","Type":"ContainerDied","Data":"fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1"} Dec 04 14:33:10 crc kubenswrapper[4848]: I1204 14:33:10.659036 4848 generic.go:334] "Generic (PLEG): container finished" podID="26bd9a14-0381-4383-a034-756b5be0e111" containerID="24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c" exitCode=0 Dec 04 14:33:10 crc kubenswrapper[4848]: I1204 14:33:10.659116 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjjxj" event={"ID":"26bd9a14-0381-4383-a034-756b5be0e111","Type":"ContainerDied","Data":"24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c"} Dec 04 14:33:11 crc kubenswrapper[4848]: I1204 14:33:11.675672 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjjxj" event={"ID":"26bd9a14-0381-4383-a034-756b5be0e111","Type":"ContainerStarted","Data":"2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041"} Dec 04 14:33:11 crc kubenswrapper[4848]: I1204 14:33:11.703995 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mjjxj" podStartSLOduration=3.09030002 podStartE2EDuration="5.703976207s" podCreationTimestamp="2025-12-04 14:33:06 +0000 UTC" firstStartedPulling="2025-12-04 14:33:08.638543597 +0000 UTC m=+2692.581040125" lastFinishedPulling="2025-12-04 14:33:11.252219784 +0000 UTC m=+2695.194716312" observedRunningTime="2025-12-04 14:33:11.696672249 +0000 UTC m=+2695.639168777" watchObservedRunningTime="2025-12-04 14:33:11.703976207 +0000 UTC m=+2695.646472735" Dec 04 14:33:16 crc kubenswrapper[4848]: I1204 14:33:16.425014 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:16 crc kubenswrapper[4848]: I1204 14:33:16.425904 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:16 crc kubenswrapper[4848]: I1204 14:33:16.452134 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:16 crc kubenswrapper[4848]: I1204 14:33:16.780097 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:16 crc kubenswrapper[4848]: I1204 14:33:16.835476 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjjxj"] Dec 04 14:33:18 crc kubenswrapper[4848]: I1204 14:33:18.750891 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mjjxj" podUID="26bd9a14-0381-4383-a034-756b5be0e111" containerName="registry-server" containerID="cri-o://2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041" gracePeriod=2 Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.311732 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.409523 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-utilities\") pod \"26bd9a14-0381-4383-a034-756b5be0e111\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.409758 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-catalog-content\") pod \"26bd9a14-0381-4383-a034-756b5be0e111\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.409819 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brz45\" (UniqueName: \"kubernetes.io/projected/26bd9a14-0381-4383-a034-756b5be0e111-kube-api-access-brz45\") pod \"26bd9a14-0381-4383-a034-756b5be0e111\" (UID: \"26bd9a14-0381-4383-a034-756b5be0e111\") " Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.411244 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-utilities" (OuterVolumeSpecName: "utilities") pod "26bd9a14-0381-4383-a034-756b5be0e111" (UID: "26bd9a14-0381-4383-a034-756b5be0e111"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.417532 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26bd9a14-0381-4383-a034-756b5be0e111-kube-api-access-brz45" (OuterVolumeSpecName: "kube-api-access-brz45") pod "26bd9a14-0381-4383-a034-756b5be0e111" (UID: "26bd9a14-0381-4383-a034-756b5be0e111"). InnerVolumeSpecName "kube-api-access-brz45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.433291 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26bd9a14-0381-4383-a034-756b5be0e111" (UID: "26bd9a14-0381-4383-a034-756b5be0e111"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.513419 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.513678 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brz45\" (UniqueName: \"kubernetes.io/projected/26bd9a14-0381-4383-a034-756b5be0e111-kube-api-access-brz45\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.513767 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26bd9a14-0381-4383-a034-756b5be0e111-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.767479 4848 generic.go:334] "Generic (PLEG): container finished" podID="26bd9a14-0381-4383-a034-756b5be0e111" containerID="2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041" exitCode=0 Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.767559 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mjjxj" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.767576 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjjxj" event={"ID":"26bd9a14-0381-4383-a034-756b5be0e111","Type":"ContainerDied","Data":"2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041"} Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.767890 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjjxj" event={"ID":"26bd9a14-0381-4383-a034-756b5be0e111","Type":"ContainerDied","Data":"086c76789e6e84fa612057ec1047e311a3501558a99cf0fa6462ddcc3302046c"} Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.767915 4848 scope.go:117] "RemoveContainer" containerID="2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.803056 4848 scope.go:117] "RemoveContainer" containerID="24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.813333 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjjxj"] Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.826509 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjjxj"] Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.829545 4848 scope.go:117] "RemoveContainer" containerID="fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.882487 4848 scope.go:117] "RemoveContainer" containerID="2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041" Dec 04 14:33:19 crc kubenswrapper[4848]: E1204 14:33:19.883173 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041\": container with ID starting with 2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041 not found: ID does not exist" containerID="2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.883221 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041"} err="failed to get container status \"2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041\": rpc error: code = NotFound desc = could not find container \"2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041\": container with ID starting with 2e94d4dc4f9feb8510141900238685538a2789ac6f248706d5ba71b4a5db0041 not found: ID does not exist" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.883248 4848 scope.go:117] "RemoveContainer" containerID="24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c" Dec 04 14:33:19 crc kubenswrapper[4848]: E1204 14:33:19.883764 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c\": container with ID starting with 24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c not found: ID does not exist" containerID="24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.883828 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c"} err="failed to get container status \"24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c\": rpc error: code = NotFound desc = could not find container \"24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c\": container with ID starting with 24cce1d7748ed2e61815400ea32e22f40a69946b5992496f2e59213433cc028c not found: ID does not exist" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.883859 4848 scope.go:117] "RemoveContainer" containerID="fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1" Dec 04 14:33:19 crc kubenswrapper[4848]: E1204 14:33:19.884324 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1\": container with ID starting with fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1 not found: ID does not exist" containerID="fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1" Dec 04 14:33:19 crc kubenswrapper[4848]: I1204 14:33:19.884383 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1"} err="failed to get container status \"fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1\": rpc error: code = NotFound desc = could not find container \"fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1\": container with ID starting with fae9e2c31d9c458adb5aded0c0501361f6514c9a9f1e974f23dd0fedbe72eae1 not found: ID does not exist" Dec 04 14:33:20 crc kubenswrapper[4848]: I1204 14:33:20.406019 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26bd9a14-0381-4383-a034-756b5be0e111" path="/var/lib/kubelet/pods/26bd9a14-0381-4383-a034-756b5be0e111/volumes" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.591348 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qs4dh"] Dec 04 14:34:41 crc kubenswrapper[4848]: E1204 14:34:41.592494 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bd9a14-0381-4383-a034-756b5be0e111" containerName="registry-server" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.592521 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bd9a14-0381-4383-a034-756b5be0e111" containerName="registry-server" Dec 04 14:34:41 crc kubenswrapper[4848]: E1204 14:34:41.592568 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bd9a14-0381-4383-a034-756b5be0e111" containerName="extract-content" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.592576 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bd9a14-0381-4383-a034-756b5be0e111" containerName="extract-content" Dec 04 14:34:41 crc kubenswrapper[4848]: E1204 14:34:41.592595 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bd9a14-0381-4383-a034-756b5be0e111" containerName="extract-utilities" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.592603 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bd9a14-0381-4383-a034-756b5be0e111" containerName="extract-utilities" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.592818 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="26bd9a14-0381-4383-a034-756b5be0e111" containerName="registry-server" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.594678 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.616814 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qs4dh"] Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.640114 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-utilities\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.640256 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcv5x\" (UniqueName: \"kubernetes.io/projected/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-kube-api-access-wcv5x\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.640277 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-catalog-content\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.743314 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcv5x\" (UniqueName: \"kubernetes.io/projected/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-kube-api-access-wcv5x\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.743866 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-catalog-content\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.744444 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-catalog-content\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.744718 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-utilities\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.745114 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-utilities\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.763834 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcv5x\" (UniqueName: \"kubernetes.io/projected/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-kube-api-access-wcv5x\") pod \"redhat-operators-qs4dh\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:41 crc kubenswrapper[4848]: I1204 14:34:41.914507 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:42 crc kubenswrapper[4848]: I1204 14:34:42.414884 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qs4dh"] Dec 04 14:34:42 crc kubenswrapper[4848]: I1204 14:34:42.675059 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs4dh" event={"ID":"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f","Type":"ContainerStarted","Data":"0baaa16e80203c7dd2d7dc0d56c566e617ad52bf26657aa3d8cd606448e013ad"} Dec 04 14:34:43 crc kubenswrapper[4848]: I1204 14:34:43.688499 4848 generic.go:334] "Generic (PLEG): container finished" podID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerID="4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad" exitCode=0 Dec 04 14:34:43 crc kubenswrapper[4848]: I1204 14:34:43.688617 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs4dh" event={"ID":"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f","Type":"ContainerDied","Data":"4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad"} Dec 04 14:34:44 crc kubenswrapper[4848]: I1204 14:34:44.313936 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:34:44 crc kubenswrapper[4848]: I1204 14:34:44.314021 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:34:44 crc kubenswrapper[4848]: I1204 14:34:44.700308 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs4dh" event={"ID":"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f","Type":"ContainerStarted","Data":"0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b"} Dec 04 14:34:49 crc kubenswrapper[4848]: I1204 14:34:49.751578 4848 generic.go:334] "Generic (PLEG): container finished" podID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerID="0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b" exitCode=0 Dec 04 14:34:49 crc kubenswrapper[4848]: I1204 14:34:49.751654 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs4dh" event={"ID":"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f","Type":"ContainerDied","Data":"0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b"} Dec 04 14:34:50 crc kubenswrapper[4848]: I1204 14:34:50.763917 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs4dh" event={"ID":"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f","Type":"ContainerStarted","Data":"a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e"} Dec 04 14:34:50 crc kubenswrapper[4848]: I1204 14:34:50.795543 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qs4dh" podStartSLOduration=3.346813837 podStartE2EDuration="9.795515459s" podCreationTimestamp="2025-12-04 14:34:41 +0000 UTC" firstStartedPulling="2025-12-04 14:34:43.690968209 +0000 UTC m=+2787.633464767" lastFinishedPulling="2025-12-04 14:34:50.139669861 +0000 UTC m=+2794.082166389" observedRunningTime="2025-12-04 14:34:50.783365903 +0000 UTC m=+2794.725862441" watchObservedRunningTime="2025-12-04 14:34:50.795515459 +0000 UTC m=+2794.738011987" Dec 04 14:34:51 crc kubenswrapper[4848]: I1204 14:34:51.915515 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:51 crc kubenswrapper[4848]: I1204 14:34:51.915885 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:34:52 crc kubenswrapper[4848]: I1204 14:34:52.970377 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qs4dh" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="registry-server" probeResult="failure" output=< Dec 04 14:34:52 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 14:34:52 crc kubenswrapper[4848]: > Dec 04 14:35:01 crc kubenswrapper[4848]: I1204 14:35:01.972563 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:35:02 crc kubenswrapper[4848]: I1204 14:35:02.030831 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:35:02 crc kubenswrapper[4848]: I1204 14:35:02.214160 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qs4dh"] Dec 04 14:35:03 crc kubenswrapper[4848]: I1204 14:35:03.908093 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qs4dh" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="registry-server" containerID="cri-o://a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e" gracePeriod=2 Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.539094 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.614912 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-catalog-content\") pod \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.615047 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-utilities\") pod \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.615100 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcv5x\" (UniqueName: \"kubernetes.io/projected/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-kube-api-access-wcv5x\") pod \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\" (UID: \"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f\") " Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.616259 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-utilities" (OuterVolumeSpecName: "utilities") pod "c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" (UID: "c2c1a8d3-303c-4ec5-9a90-03ca0c95411f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.629154 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-kube-api-access-wcv5x" (OuterVolumeSpecName: "kube-api-access-wcv5x") pod "c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" (UID: "c2c1a8d3-303c-4ec5-9a90-03ca0c95411f"). InnerVolumeSpecName "kube-api-access-wcv5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.718277 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.718309 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcv5x\" (UniqueName: \"kubernetes.io/projected/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-kube-api-access-wcv5x\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.754586 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" (UID: "c2c1a8d3-303c-4ec5-9a90-03ca0c95411f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.820779 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.927558 4848 generic.go:334] "Generic (PLEG): container finished" podID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerID="a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e" exitCode=0 Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.927654 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs4dh" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.927675 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs4dh" event={"ID":"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f","Type":"ContainerDied","Data":"a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e"} Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.928671 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs4dh" event={"ID":"c2c1a8d3-303c-4ec5-9a90-03ca0c95411f","Type":"ContainerDied","Data":"0baaa16e80203c7dd2d7dc0d56c566e617ad52bf26657aa3d8cd606448e013ad"} Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.928715 4848 scope.go:117] "RemoveContainer" containerID="a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.966667 4848 scope.go:117] "RemoveContainer" containerID="0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b" Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.968385 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qs4dh"] Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.979121 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qs4dh"] Dec 04 14:35:04 crc kubenswrapper[4848]: I1204 14:35:04.995235 4848 scope.go:117] "RemoveContainer" containerID="4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad" Dec 04 14:35:05 crc kubenswrapper[4848]: I1204 14:35:05.038526 4848 scope.go:117] "RemoveContainer" containerID="a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e" Dec 04 14:35:05 crc kubenswrapper[4848]: E1204 14:35:05.039138 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e\": container with ID starting with a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e not found: ID does not exist" containerID="a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e" Dec 04 14:35:05 crc kubenswrapper[4848]: I1204 14:35:05.039174 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e"} err="failed to get container status \"a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e\": rpc error: code = NotFound desc = could not find container \"a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e\": container with ID starting with a3b07e7d84a90b156d47500eb1a95843eb888c35643365951ed66a01fd41b00e not found: ID does not exist" Dec 04 14:35:05 crc kubenswrapper[4848]: I1204 14:35:05.039203 4848 scope.go:117] "RemoveContainer" containerID="0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b" Dec 04 14:35:05 crc kubenswrapper[4848]: E1204 14:35:05.039583 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b\": container with ID starting with 0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b not found: ID does not exist" containerID="0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b" Dec 04 14:35:05 crc kubenswrapper[4848]: I1204 14:35:05.039628 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b"} err="failed to get container status \"0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b\": rpc error: code = NotFound desc = could not find container \"0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b\": container with ID starting with 0c668b19ed72d41297ca3bca150c0342a1ebfae51f3541b799385002c2689b3b not found: ID does not exist" Dec 04 14:35:05 crc kubenswrapper[4848]: I1204 14:35:05.039659 4848 scope.go:117] "RemoveContainer" containerID="4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad" Dec 04 14:35:05 crc kubenswrapper[4848]: E1204 14:35:05.040242 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad\": container with ID starting with 4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad not found: ID does not exist" containerID="4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad" Dec 04 14:35:05 crc kubenswrapper[4848]: I1204 14:35:05.040301 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad"} err="failed to get container status \"4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad\": rpc error: code = NotFound desc = could not find container \"4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad\": container with ID starting with 4614f36328388339b602525317123fd136090d403feb93a75c2811777e6845ad not found: ID does not exist" Dec 04 14:35:06 crc kubenswrapper[4848]: I1204 14:35:06.406778 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" path="/var/lib/kubelet/pods/c2c1a8d3-303c-4ec5-9a90-03ca0c95411f/volumes" Dec 04 14:35:14 crc kubenswrapper[4848]: I1204 14:35:14.313884 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:35:14 crc kubenswrapper[4848]: I1204 14:35:14.314502 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:35:44 crc kubenswrapper[4848]: I1204 14:35:44.314192 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:35:44 crc kubenswrapper[4848]: I1204 14:35:44.314777 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:35:44 crc kubenswrapper[4848]: I1204 14:35:44.314832 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:35:44 crc kubenswrapper[4848]: I1204 14:35:44.315812 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef540d9e8a947299436fb6249a1ab4f19679231bf019f3be88d55d65c9bd9249"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:35:44 crc kubenswrapper[4848]: I1204 14:35:44.315868 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://ef540d9e8a947299436fb6249a1ab4f19679231bf019f3be88d55d65c9bd9249" gracePeriod=600 Dec 04 14:35:44 crc kubenswrapper[4848]: I1204 14:35:44.495856 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="ef540d9e8a947299436fb6249a1ab4f19679231bf019f3be88d55d65c9bd9249" exitCode=0 Dec 04 14:35:44 crc kubenswrapper[4848]: I1204 14:35:44.495896 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"ef540d9e8a947299436fb6249a1ab4f19679231bf019f3be88d55d65c9bd9249"} Dec 04 14:35:44 crc kubenswrapper[4848]: I1204 14:35:44.495926 4848 scope.go:117] "RemoveContainer" containerID="47c8f875c6436bf1aed6dfaf92da7fe8c611899a29c0c72f256609dcf33339d6" Dec 04 14:35:45 crc kubenswrapper[4848]: I1204 14:35:45.508704 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a"} Dec 04 14:36:01 crc kubenswrapper[4848]: I1204 14:36:01.660452 4848 generic.go:334] "Generic (PLEG): container finished" podID="3146cd9a-c65d-4942-bb4d-ee0951403d87" containerID="067d3258f855f684d35aa34fc141f010e30a9cd19fa44ff7c97397c6e4d9a44e" exitCode=0 Dec 04 14:36:01 crc kubenswrapper[4848]: I1204 14:36:01.660558 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" event={"ID":"3146cd9a-c65d-4942-bb4d-ee0951403d87","Type":"ContainerDied","Data":"067d3258f855f684d35aa34fc141f010e30a9cd19fa44ff7c97397c6e4d9a44e"} Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.181253 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.302058 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-secret-0\") pod \"3146cd9a-c65d-4942-bb4d-ee0951403d87\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.302221 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-ssh-key\") pod \"3146cd9a-c65d-4942-bb4d-ee0951403d87\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.302299 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p489f\" (UniqueName: \"kubernetes.io/projected/3146cd9a-c65d-4942-bb4d-ee0951403d87-kube-api-access-p489f\") pod \"3146cd9a-c65d-4942-bb4d-ee0951403d87\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.302343 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-inventory\") pod \"3146cd9a-c65d-4942-bb4d-ee0951403d87\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.302399 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-combined-ca-bundle\") pod \"3146cd9a-c65d-4942-bb4d-ee0951403d87\" (UID: \"3146cd9a-c65d-4942-bb4d-ee0951403d87\") " Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.329700 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3146cd9a-c65d-4942-bb4d-ee0951403d87" (UID: "3146cd9a-c65d-4942-bb4d-ee0951403d87"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.358240 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3146cd9a-c65d-4942-bb4d-ee0951403d87-kube-api-access-p489f" (OuterVolumeSpecName: "kube-api-access-p489f") pod "3146cd9a-c65d-4942-bb4d-ee0951403d87" (UID: "3146cd9a-c65d-4942-bb4d-ee0951403d87"). InnerVolumeSpecName "kube-api-access-p489f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.400555 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "3146cd9a-c65d-4942-bb4d-ee0951403d87" (UID: "3146cd9a-c65d-4942-bb4d-ee0951403d87"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.401775 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-inventory" (OuterVolumeSpecName: "inventory") pod "3146cd9a-c65d-4942-bb4d-ee0951403d87" (UID: "3146cd9a-c65d-4942-bb4d-ee0951403d87"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.405151 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p489f\" (UniqueName: \"kubernetes.io/projected/3146cd9a-c65d-4942-bb4d-ee0951403d87-kube-api-access-p489f\") on node \"crc\" DevicePath \"\"" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.405188 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.405198 4848 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.405208 4848 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.415470 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3146cd9a-c65d-4942-bb4d-ee0951403d87" (UID: "3146cd9a-c65d-4942-bb4d-ee0951403d87"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.537851 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3146cd9a-c65d-4942-bb4d-ee0951403d87-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.694327 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" event={"ID":"3146cd9a-c65d-4942-bb4d-ee0951403d87","Type":"ContainerDied","Data":"e8feb2ca652ff2d7806fb7892c2a0b3498822424a1308da9d355bbb2ee2933f9"} Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.694373 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8feb2ca652ff2d7806fb7892c2a0b3498822424a1308da9d355bbb2ee2933f9" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.694445 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6lph2" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.788379 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp"] Dec 04 14:36:03 crc kubenswrapper[4848]: E1204 14:36:03.788998 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="registry-server" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.789018 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="registry-server" Dec 04 14:36:03 crc kubenswrapper[4848]: E1204 14:36:03.789035 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="extract-utilities" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.789042 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="extract-utilities" Dec 04 14:36:03 crc kubenswrapper[4848]: E1204 14:36:03.790014 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="extract-content" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.790035 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="extract-content" Dec 04 14:36:03 crc kubenswrapper[4848]: E1204 14:36:03.790076 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3146cd9a-c65d-4942-bb4d-ee0951403d87" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.790086 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="3146cd9a-c65d-4942-bb4d-ee0951403d87" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.790477 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c1a8d3-303c-4ec5-9a90-03ca0c95411f" containerName="registry-server" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.790531 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="3146cd9a-c65d-4942-bb4d-ee0951403d87" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.792185 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.796263 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.796561 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.796908 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.797310 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.797485 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.797614 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.797734 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.817468 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp"] Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.954734 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.954833 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.954902 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.955063 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.955464 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.955820 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-582vc\" (UniqueName: \"kubernetes.io/projected/86fbd978-880e-415b-ad1e-e367a0c0079b-kube-api-access-582vc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.955869 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.956076 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:03 crc kubenswrapper[4848]: I1204 14:36:03.956223 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.058662 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.058733 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.058822 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.058899 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.059042 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.059128 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-582vc\" (UniqueName: \"kubernetes.io/projected/86fbd978-880e-415b-ad1e-e367a0c0079b-kube-api-access-582vc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.059159 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.059198 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.059248 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.061034 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.062623 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.062890 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.063433 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.064628 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.065377 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.066001 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.067477 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.079725 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-582vc\" (UniqueName: \"kubernetes.io/projected/86fbd978-880e-415b-ad1e-e367a0c0079b-kube-api-access-582vc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-v9fxp\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.113700 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:36:04 crc kubenswrapper[4848]: I1204 14:36:04.725192 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp"] Dec 04 14:36:04 crc kubenswrapper[4848]: W1204 14:36:04.735089 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86fbd978_880e_415b_ad1e_e367a0c0079b.slice/crio-7dff83e1142eba3d5f6788379407347a385b1505028ce7f69a573f60e913f437 WatchSource:0}: Error finding container 7dff83e1142eba3d5f6788379407347a385b1505028ce7f69a573f60e913f437: Status 404 returned error can't find the container with id 7dff83e1142eba3d5f6788379407347a385b1505028ce7f69a573f60e913f437 Dec 04 14:36:05 crc kubenswrapper[4848]: I1204 14:36:05.714893 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" event={"ID":"86fbd978-880e-415b-ad1e-e367a0c0079b","Type":"ContainerStarted","Data":"b36d2a3b41a203d3fa4255e901fe3458f3449bf412a27e9308bae52366a4aaa9"} Dec 04 14:36:05 crc kubenswrapper[4848]: I1204 14:36:05.715414 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" event={"ID":"86fbd978-880e-415b-ad1e-e367a0c0079b","Type":"ContainerStarted","Data":"7dff83e1142eba3d5f6788379407347a385b1505028ce7f69a573f60e913f437"} Dec 04 14:36:05 crc kubenswrapper[4848]: I1204 14:36:05.741996 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" podStartSLOduration=2.322260533 podStartE2EDuration="2.741975332s" podCreationTimestamp="2025-12-04 14:36:03 +0000 UTC" firstStartedPulling="2025-12-04 14:36:04.737408535 +0000 UTC m=+2868.679905063" lastFinishedPulling="2025-12-04 14:36:05.157123334 +0000 UTC m=+2869.099619862" observedRunningTime="2025-12-04 14:36:05.733170988 +0000 UTC m=+2869.675667526" watchObservedRunningTime="2025-12-04 14:36:05.741975332 +0000 UTC m=+2869.684471860" Dec 04 14:37:44 crc kubenswrapper[4848]: I1204 14:37:44.314283 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:37:44 crc kubenswrapper[4848]: I1204 14:37:44.314794 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:38:14 crc kubenswrapper[4848]: I1204 14:38:14.314250 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:38:14 crc kubenswrapper[4848]: I1204 14:38:14.314834 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:38:44 crc kubenswrapper[4848]: I1204 14:38:44.314207 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:38:44 crc kubenswrapper[4848]: I1204 14:38:44.314828 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:38:44 crc kubenswrapper[4848]: I1204 14:38:44.314880 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:38:44 crc kubenswrapper[4848]: I1204 14:38:44.315895 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:38:44 crc kubenswrapper[4848]: I1204 14:38:44.316023 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" gracePeriod=600 Dec 04 14:38:44 crc kubenswrapper[4848]: E1204 14:38:44.449258 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:38:45 crc kubenswrapper[4848]: I1204 14:38:45.449657 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" exitCode=0 Dec 04 14:38:45 crc kubenswrapper[4848]: I1204 14:38:45.449874 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a"} Dec 04 14:38:45 crc kubenswrapper[4848]: I1204 14:38:45.450027 4848 scope.go:117] "RemoveContainer" containerID="ef540d9e8a947299436fb6249a1ab4f19679231bf019f3be88d55d65c9bd9249" Dec 04 14:38:45 crc kubenswrapper[4848]: I1204 14:38:45.451541 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:38:45 crc kubenswrapper[4848]: E1204 14:38:45.451840 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:38:49 crc kubenswrapper[4848]: I1204 14:38:49.499545 4848 generic.go:334] "Generic (PLEG): container finished" podID="86fbd978-880e-415b-ad1e-e367a0c0079b" containerID="b36d2a3b41a203d3fa4255e901fe3458f3449bf412a27e9308bae52366a4aaa9" exitCode=0 Dec 04 14:38:49 crc kubenswrapper[4848]: I1204 14:38:49.499659 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" event={"ID":"86fbd978-880e-415b-ad1e-e367a0c0079b","Type":"ContainerDied","Data":"b36d2a3b41a203d3fa4255e901fe3458f3449bf412a27e9308bae52366a4aaa9"} Dec 04 14:38:50 crc kubenswrapper[4848]: I1204 14:38:50.999139 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.126910 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-extra-config-0\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.126990 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-ssh-key\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.127093 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-0\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.127115 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-0\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.127165 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-1\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.127219 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-inventory\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.127735 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-1\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.127758 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-combined-ca-bundle\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.127836 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-582vc\" (UniqueName: \"kubernetes.io/projected/86fbd978-880e-415b-ad1e-e367a0c0079b-kube-api-access-582vc\") pod \"86fbd978-880e-415b-ad1e-e367a0c0079b\" (UID: \"86fbd978-880e-415b-ad1e-e367a0c0079b\") " Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.132388 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86fbd978-880e-415b-ad1e-e367a0c0079b-kube-api-access-582vc" (OuterVolumeSpecName: "kube-api-access-582vc") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "kube-api-access-582vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.137660 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.158213 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.164143 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.165395 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.168044 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.168591 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.169150 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-inventory" (OuterVolumeSpecName: "inventory") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.170472 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "86fbd978-880e-415b-ad1e-e367a0c0079b" (UID: "86fbd978-880e-415b-ad1e-e367a0c0079b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231581 4848 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231616 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231626 4848 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231639 4848 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231648 4848 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231656 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231664 4848 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231673 4848 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fbd978-880e-415b-ad1e-e367a0c0079b-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.231683 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-582vc\" (UniqueName: \"kubernetes.io/projected/86fbd978-880e-415b-ad1e-e367a0c0079b-kube-api-access-582vc\") on node \"crc\" DevicePath \"\"" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.522119 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" event={"ID":"86fbd978-880e-415b-ad1e-e367a0c0079b","Type":"ContainerDied","Data":"7dff83e1142eba3d5f6788379407347a385b1505028ce7f69a573f60e913f437"} Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.522159 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dff83e1142eba3d5f6788379407347a385b1505028ce7f69a573f60e913f437" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.522177 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-v9fxp" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.633304 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs"] Dec 04 14:38:51 crc kubenswrapper[4848]: E1204 14:38:51.633867 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86fbd978-880e-415b-ad1e-e367a0c0079b" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.633883 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="86fbd978-880e-415b-ad1e-e367a0c0079b" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.634118 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="86fbd978-880e-415b-ad1e-e367a0c0079b" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.635023 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.637536 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.637570 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.637689 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.638001 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.638058 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.647411 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs"] Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.743366 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.743615 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.743735 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.743823 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdft9\" (UniqueName: \"kubernetes.io/projected/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-kube-api-access-jdft9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.743884 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.744129 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.744164 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.846406 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.846478 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.846521 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdft9\" (UniqueName: \"kubernetes.io/projected/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-kube-api-access-jdft9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.846540 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.846627 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.846652 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.846705 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.851389 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.852054 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.852489 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.852916 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.853197 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.855217 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.864639 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdft9\" (UniqueName: \"kubernetes.io/projected/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-kube-api-access-jdft9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:51 crc kubenswrapper[4848]: I1204 14:38:51.960392 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:38:52 crc kubenswrapper[4848]: I1204 14:38:52.518786 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs"] Dec 04 14:38:52 crc kubenswrapper[4848]: I1204 14:38:52.534452 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:38:52 crc kubenswrapper[4848]: I1204 14:38:52.543109 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" event={"ID":"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1","Type":"ContainerStarted","Data":"1585f82a13b96bd8291ada31515e1a11b3ee91ea864c36524495a41932ae178b"} Dec 04 14:38:53 crc kubenswrapper[4848]: I1204 14:38:53.553236 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" event={"ID":"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1","Type":"ContainerStarted","Data":"8c41002cdb2a4a5713c89816d7fbe5630f84350adbbe9a4689c2605e944ab76c"} Dec 04 14:38:53 crc kubenswrapper[4848]: I1204 14:38:53.580192 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" podStartSLOduration=1.9903329140000001 podStartE2EDuration="2.580165274s" podCreationTimestamp="2025-12-04 14:38:51 +0000 UTC" firstStartedPulling="2025-12-04 14:38:52.534240099 +0000 UTC m=+3036.476736617" lastFinishedPulling="2025-12-04 14:38:53.124072429 +0000 UTC m=+3037.066568977" observedRunningTime="2025-12-04 14:38:53.567328331 +0000 UTC m=+3037.509824859" watchObservedRunningTime="2025-12-04 14:38:53.580165274 +0000 UTC m=+3037.522661802" Dec 04 14:38:58 crc kubenswrapper[4848]: I1204 14:38:58.393189 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:38:58 crc kubenswrapper[4848]: E1204 14:38:58.395388 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:39:13 crc kubenswrapper[4848]: I1204 14:39:13.394042 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:39:13 crc kubenswrapper[4848]: E1204 14:39:13.394816 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:39:25 crc kubenswrapper[4848]: I1204 14:39:25.393171 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:39:25 crc kubenswrapper[4848]: E1204 14:39:25.395022 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:39:36 crc kubenswrapper[4848]: I1204 14:39:36.403126 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:39:36 crc kubenswrapper[4848]: E1204 14:39:36.403883 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:39:47 crc kubenswrapper[4848]: I1204 14:39:47.393328 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:39:47 crc kubenswrapper[4848]: E1204 14:39:47.394174 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:39:59 crc kubenswrapper[4848]: I1204 14:39:59.395108 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:39:59 crc kubenswrapper[4848]: E1204 14:39:59.395758 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:40:14 crc kubenswrapper[4848]: I1204 14:40:14.393780 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:40:14 crc kubenswrapper[4848]: E1204 14:40:14.394613 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:40:28 crc kubenswrapper[4848]: I1204 14:40:28.394255 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:40:28 crc kubenswrapper[4848]: E1204 14:40:28.395146 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:40:43 crc kubenswrapper[4848]: I1204 14:40:43.397344 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:40:43 crc kubenswrapper[4848]: E1204 14:40:43.399692 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:40:57 crc kubenswrapper[4848]: I1204 14:40:57.393221 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:40:57 crc kubenswrapper[4848]: E1204 14:40:57.393974 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:41:10 crc kubenswrapper[4848]: I1204 14:41:10.393701 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:41:10 crc kubenswrapper[4848]: E1204 14:41:10.394541 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:41:13 crc kubenswrapper[4848]: I1204 14:41:13.759081 4848 generic.go:334] "Generic (PLEG): container finished" podID="d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" containerID="8c41002cdb2a4a5713c89816d7fbe5630f84350adbbe9a4689c2605e944ab76c" exitCode=0 Dec 04 14:41:13 crc kubenswrapper[4848]: I1204 14:41:13.759162 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" event={"ID":"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1","Type":"ContainerDied","Data":"8c41002cdb2a4a5713c89816d7fbe5630f84350adbbe9a4689c2605e944ab76c"} Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.230136 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.384317 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdft9\" (UniqueName: \"kubernetes.io/projected/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-kube-api-access-jdft9\") pod \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.385118 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-inventory\") pod \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.385263 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ssh-key\") pod \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.385500 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-0\") pod \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.385755 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-1\") pod \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.385907 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-2\") pod \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.386080 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-telemetry-combined-ca-bundle\") pod \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\" (UID: \"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1\") " Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.392591 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" (UID: "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.394221 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-kube-api-access-jdft9" (OuterVolumeSpecName: "kube-api-access-jdft9") pod "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" (UID: "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1"). InnerVolumeSpecName "kube-api-access-jdft9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.418607 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" (UID: "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.421928 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" (UID: "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.425431 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" (UID: "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.428074 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-inventory" (OuterVolumeSpecName: "inventory") pod "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" (UID: "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.438345 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" (UID: "d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.488591 4848 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.488641 4848 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.488658 4848 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.488674 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdft9\" (UniqueName: \"kubernetes.io/projected/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-kube-api-access-jdft9\") on node \"crc\" DevicePath \"\"" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.488687 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.488696 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.488708 4848 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.783228 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" event={"ID":"d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1","Type":"ContainerDied","Data":"1585f82a13b96bd8291ada31515e1a11b3ee91ea864c36524495a41932ae178b"} Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.783291 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1585f82a13b96bd8291ada31515e1a11b3ee91ea864c36524495a41932ae178b" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.783374 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.872444 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg"] Dec 04 14:41:15 crc kubenswrapper[4848]: E1204 14:41:15.873087 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.873111 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.873420 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.874480 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.876428 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.876891 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.877061 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.877410 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.877410 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.879990 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg"] Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.899103 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.899410 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dflwr\" (UniqueName: \"kubernetes.io/projected/0e112a8b-6965-4e32-bb39-4ff08519bdee-kube-api-access-dflwr\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.899561 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.899644 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.899741 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.899786 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:15 crc kubenswrapper[4848]: I1204 14:41:15.899844 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.002773 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.002890 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.003013 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dflwr\" (UniqueName: \"kubernetes.io/projected/0e112a8b-6965-4e32-bb39-4ff08519bdee-kube-api-access-dflwr\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.003069 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.003099 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.003163 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.003206 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.007360 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.007360 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.007792 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.008629 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.009941 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.010256 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.020312 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dflwr\" (UniqueName: \"kubernetes.io/projected/0e112a8b-6965-4e32-bb39-4ff08519bdee-kube-api-access-dflwr\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.202798 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.723420 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg"] Dec 04 14:41:16 crc kubenswrapper[4848]: I1204 14:41:16.794460 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" event={"ID":"0e112a8b-6965-4e32-bb39-4ff08519bdee","Type":"ContainerStarted","Data":"bef7de153dc8ea34e573c0cdbac5258398e4eda4bf0b59e92ba593d55dab3d9d"} Dec 04 14:41:17 crc kubenswrapper[4848]: I1204 14:41:17.196897 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:41:17 crc kubenswrapper[4848]: I1204 14:41:17.804439 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" event={"ID":"0e112a8b-6965-4e32-bb39-4ff08519bdee","Type":"ContainerStarted","Data":"3eac584ded7accd4bec2bbb213e969f1a5b08765b90e12941c4bc41c242fd819"} Dec 04 14:41:17 crc kubenswrapper[4848]: I1204 14:41:17.826353 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" podStartSLOduration=2.366505811 podStartE2EDuration="2.826330639s" podCreationTimestamp="2025-12-04 14:41:15 +0000 UTC" firstStartedPulling="2025-12-04 14:41:16.734066038 +0000 UTC m=+3180.676562586" lastFinishedPulling="2025-12-04 14:41:17.193890876 +0000 UTC m=+3181.136387414" observedRunningTime="2025-12-04 14:41:17.819707029 +0000 UTC m=+3181.762203557" watchObservedRunningTime="2025-12-04 14:41:17.826330639 +0000 UTC m=+3181.768827167" Dec 04 14:41:23 crc kubenswrapper[4848]: I1204 14:41:23.395199 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:41:23 crc kubenswrapper[4848]: E1204 14:41:23.397478 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:41:38 crc kubenswrapper[4848]: I1204 14:41:38.393764 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:41:38 crc kubenswrapper[4848]: E1204 14:41:38.394574 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:41:51 crc kubenswrapper[4848]: I1204 14:41:51.394425 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:41:51 crc kubenswrapper[4848]: E1204 14:41:51.396542 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:42:06 crc kubenswrapper[4848]: I1204 14:42:06.401565 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:42:06 crc kubenswrapper[4848]: E1204 14:42:06.402517 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:42:21 crc kubenswrapper[4848]: I1204 14:42:21.393772 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:42:21 crc kubenswrapper[4848]: E1204 14:42:21.394560 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.289082 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mj6f6"] Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.291670 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.309078 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mj6f6"] Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.365475 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-catalog-content\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.365736 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csc69\" (UniqueName: \"kubernetes.io/projected/27e25768-61f4-4749-9a5c-006da09aff30-kube-api-access-csc69\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.365831 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-utilities\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.467862 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-catalog-content\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.468432 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csc69\" (UniqueName: \"kubernetes.io/projected/27e25768-61f4-4749-9a5c-006da09aff30-kube-api-access-csc69\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.468556 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-utilities\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.469170 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-catalog-content\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.469199 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-utilities\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.486097 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csc69\" (UniqueName: \"kubernetes.io/projected/27e25768-61f4-4749-9a5c-006da09aff30-kube-api-access-csc69\") pod \"community-operators-mj6f6\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:22 crc kubenswrapper[4848]: I1204 14:42:22.616265 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:23 crc kubenswrapper[4848]: I1204 14:42:23.187288 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mj6f6"] Dec 04 14:42:23 crc kubenswrapper[4848]: I1204 14:42:23.547736 4848 generic.go:334] "Generic (PLEG): container finished" podID="27e25768-61f4-4749-9a5c-006da09aff30" containerID="f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286" exitCode=0 Dec 04 14:42:23 crc kubenswrapper[4848]: I1204 14:42:23.547814 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mj6f6" event={"ID":"27e25768-61f4-4749-9a5c-006da09aff30","Type":"ContainerDied","Data":"f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286"} Dec 04 14:42:23 crc kubenswrapper[4848]: I1204 14:42:23.548638 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mj6f6" event={"ID":"27e25768-61f4-4749-9a5c-006da09aff30","Type":"ContainerStarted","Data":"f7f11cd42491995d3a1db5d37ac3e5d7c2f9c6acf79e6d65da7259de4de1d8d2"} Dec 04 14:42:24 crc kubenswrapper[4848]: I1204 14:42:24.560512 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mj6f6" event={"ID":"27e25768-61f4-4749-9a5c-006da09aff30","Type":"ContainerStarted","Data":"1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648"} Dec 04 14:42:25 crc kubenswrapper[4848]: I1204 14:42:25.571896 4848 generic.go:334] "Generic (PLEG): container finished" podID="27e25768-61f4-4749-9a5c-006da09aff30" containerID="1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648" exitCode=0 Dec 04 14:42:25 crc kubenswrapper[4848]: I1204 14:42:25.571974 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mj6f6" event={"ID":"27e25768-61f4-4749-9a5c-006da09aff30","Type":"ContainerDied","Data":"1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648"} Dec 04 14:42:27 crc kubenswrapper[4848]: I1204 14:42:27.596125 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mj6f6" event={"ID":"27e25768-61f4-4749-9a5c-006da09aff30","Type":"ContainerStarted","Data":"2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898"} Dec 04 14:42:27 crc kubenswrapper[4848]: I1204 14:42:27.618364 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mj6f6" podStartSLOduration=2.521466307 podStartE2EDuration="5.618333194s" podCreationTimestamp="2025-12-04 14:42:22 +0000 UTC" firstStartedPulling="2025-12-04 14:42:23.550722317 +0000 UTC m=+3247.493218845" lastFinishedPulling="2025-12-04 14:42:26.647589194 +0000 UTC m=+3250.590085732" observedRunningTime="2025-12-04 14:42:27.615540116 +0000 UTC m=+3251.558036644" watchObservedRunningTime="2025-12-04 14:42:27.618333194 +0000 UTC m=+3251.560829732" Dec 04 14:42:32 crc kubenswrapper[4848]: I1204 14:42:32.617195 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:32 crc kubenswrapper[4848]: I1204 14:42:32.617762 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:32 crc kubenswrapper[4848]: I1204 14:42:32.683231 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:32 crc kubenswrapper[4848]: I1204 14:42:32.735961 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:32 crc kubenswrapper[4848]: I1204 14:42:32.921238 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mj6f6"] Dec 04 14:42:34 crc kubenswrapper[4848]: I1204 14:42:34.688645 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mj6f6" podUID="27e25768-61f4-4749-9a5c-006da09aff30" containerName="registry-server" containerID="cri-o://2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898" gracePeriod=2 Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.276445 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.393716 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:42:35 crc kubenswrapper[4848]: E1204 14:42:35.394111 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.397129 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csc69\" (UniqueName: \"kubernetes.io/projected/27e25768-61f4-4749-9a5c-006da09aff30-kube-api-access-csc69\") pod \"27e25768-61f4-4749-9a5c-006da09aff30\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.397184 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-utilities\") pod \"27e25768-61f4-4749-9a5c-006da09aff30\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.397530 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-catalog-content\") pod \"27e25768-61f4-4749-9a5c-006da09aff30\" (UID: \"27e25768-61f4-4749-9a5c-006da09aff30\") " Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.397859 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-utilities" (OuterVolumeSpecName: "utilities") pod "27e25768-61f4-4749-9a5c-006da09aff30" (UID: "27e25768-61f4-4749-9a5c-006da09aff30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.398440 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.402452 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e25768-61f4-4749-9a5c-006da09aff30-kube-api-access-csc69" (OuterVolumeSpecName: "kube-api-access-csc69") pod "27e25768-61f4-4749-9a5c-006da09aff30" (UID: "27e25768-61f4-4749-9a5c-006da09aff30"). InnerVolumeSpecName "kube-api-access-csc69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.448908 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27e25768-61f4-4749-9a5c-006da09aff30" (UID: "27e25768-61f4-4749-9a5c-006da09aff30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.500501 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csc69\" (UniqueName: \"kubernetes.io/projected/27e25768-61f4-4749-9a5c-006da09aff30-kube-api-access-csc69\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.500543 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27e25768-61f4-4749-9a5c-006da09aff30-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.701176 4848 generic.go:334] "Generic (PLEG): container finished" podID="27e25768-61f4-4749-9a5c-006da09aff30" containerID="2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898" exitCode=0 Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.701252 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mj6f6" event={"ID":"27e25768-61f4-4749-9a5c-006da09aff30","Type":"ContainerDied","Data":"2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898"} Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.701488 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mj6f6" event={"ID":"27e25768-61f4-4749-9a5c-006da09aff30","Type":"ContainerDied","Data":"f7f11cd42491995d3a1db5d37ac3e5d7c2f9c6acf79e6d65da7259de4de1d8d2"} Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.701521 4848 scope.go:117] "RemoveContainer" containerID="2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.701324 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mj6f6" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.721807 4848 scope.go:117] "RemoveContainer" containerID="1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.754766 4848 scope.go:117] "RemoveContainer" containerID="f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.809634 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mj6f6"] Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.809845 4848 scope.go:117] "RemoveContainer" containerID="2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898" Dec 04 14:42:35 crc kubenswrapper[4848]: E1204 14:42:35.810454 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898\": container with ID starting with 2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898 not found: ID does not exist" containerID="2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.810500 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898"} err="failed to get container status \"2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898\": rpc error: code = NotFound desc = could not find container \"2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898\": container with ID starting with 2a344aaabf5dbcefde7e9e1ae028915a615e3ba0e7ddd64a953d952e04eba898 not found: ID does not exist" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.810528 4848 scope.go:117] "RemoveContainer" containerID="1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648" Dec 04 14:42:35 crc kubenswrapper[4848]: E1204 14:42:35.810905 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648\": container with ID starting with 1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648 not found: ID does not exist" containerID="1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.810939 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648"} err="failed to get container status \"1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648\": rpc error: code = NotFound desc = could not find container \"1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648\": container with ID starting with 1cb677498a5684965edaeed53bcb2723c8c60901db1ca3499a59c96df3ba2648 not found: ID does not exist" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.810974 4848 scope.go:117] "RemoveContainer" containerID="f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286" Dec 04 14:42:35 crc kubenswrapper[4848]: E1204 14:42:35.811292 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286\": container with ID starting with f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286 not found: ID does not exist" containerID="f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.811373 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286"} err="failed to get container status \"f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286\": rpc error: code = NotFound desc = could not find container \"f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286\": container with ID starting with f990b1a1b1407dde107a126efd673d42a17f3c412cddc58da9a6e2dae2003286 not found: ID does not exist" Dec 04 14:42:35 crc kubenswrapper[4848]: I1204 14:42:35.821326 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mj6f6"] Dec 04 14:42:36 crc kubenswrapper[4848]: I1204 14:42:36.406012 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27e25768-61f4-4749-9a5c-006da09aff30" path="/var/lib/kubelet/pods/27e25768-61f4-4749-9a5c-006da09aff30/volumes" Dec 04 14:42:50 crc kubenswrapper[4848]: I1204 14:42:50.395120 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:42:50 crc kubenswrapper[4848]: E1204 14:42:50.396104 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:43:01 crc kubenswrapper[4848]: I1204 14:43:01.394366 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:43:01 crc kubenswrapper[4848]: E1204 14:43:01.395514 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:43:11 crc kubenswrapper[4848]: I1204 14:43:11.074607 4848 generic.go:334] "Generic (PLEG): container finished" podID="0e112a8b-6965-4e32-bb39-4ff08519bdee" containerID="3eac584ded7accd4bec2bbb213e969f1a5b08765b90e12941c4bc41c242fd819" exitCode=0 Dec 04 14:43:11 crc kubenswrapper[4848]: I1204 14:43:11.074696 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" event={"ID":"0e112a8b-6965-4e32-bb39-4ff08519bdee","Type":"ContainerDied","Data":"3eac584ded7accd4bec2bbb213e969f1a5b08765b90e12941c4bc41c242fd819"} Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.394273 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:43:12 crc kubenswrapper[4848]: E1204 14:43:12.394768 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.547354 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.602726 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-2\") pod \"0e112a8b-6965-4e32-bb39-4ff08519bdee\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.602886 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-0\") pod \"0e112a8b-6965-4e32-bb39-4ff08519bdee\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.602909 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dflwr\" (UniqueName: \"kubernetes.io/projected/0e112a8b-6965-4e32-bb39-4ff08519bdee-kube-api-access-dflwr\") pod \"0e112a8b-6965-4e32-bb39-4ff08519bdee\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.602983 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-1\") pod \"0e112a8b-6965-4e32-bb39-4ff08519bdee\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.603098 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ssh-key\") pod \"0e112a8b-6965-4e32-bb39-4ff08519bdee\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.603146 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-inventory\") pod \"0e112a8b-6965-4e32-bb39-4ff08519bdee\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.603195 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-telemetry-power-monitoring-combined-ca-bundle\") pod \"0e112a8b-6965-4e32-bb39-4ff08519bdee\" (UID: \"0e112a8b-6965-4e32-bb39-4ff08519bdee\") " Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.610063 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "0e112a8b-6965-4e32-bb39-4ff08519bdee" (UID: "0e112a8b-6965-4e32-bb39-4ff08519bdee"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.610752 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e112a8b-6965-4e32-bb39-4ff08519bdee-kube-api-access-dflwr" (OuterVolumeSpecName: "kube-api-access-dflwr") pod "0e112a8b-6965-4e32-bb39-4ff08519bdee" (UID: "0e112a8b-6965-4e32-bb39-4ff08519bdee"). InnerVolumeSpecName "kube-api-access-dflwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.636399 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "0e112a8b-6965-4e32-bb39-4ff08519bdee" (UID: "0e112a8b-6965-4e32-bb39-4ff08519bdee"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.636745 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "0e112a8b-6965-4e32-bb39-4ff08519bdee" (UID: "0e112a8b-6965-4e32-bb39-4ff08519bdee"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.643589 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0e112a8b-6965-4e32-bb39-4ff08519bdee" (UID: "0e112a8b-6965-4e32-bb39-4ff08519bdee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.644590 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-inventory" (OuterVolumeSpecName: "inventory") pod "0e112a8b-6965-4e32-bb39-4ff08519bdee" (UID: "0e112a8b-6965-4e32-bb39-4ff08519bdee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.647375 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "0e112a8b-6965-4e32-bb39-4ff08519bdee" (UID: "0e112a8b-6965-4e32-bb39-4ff08519bdee"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.711322 4848 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.711565 4848 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.711711 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dflwr\" (UniqueName: \"kubernetes.io/projected/0e112a8b-6965-4e32-bb39-4ff08519bdee-kube-api-access-dflwr\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.711770 4848 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.711833 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.711888 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:12 crc kubenswrapper[4848]: I1204 14:43:12.711962 4848 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e112a8b-6965-4e32-bb39-4ff08519bdee-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.094879 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" event={"ID":"0e112a8b-6965-4e32-bb39-4ff08519bdee","Type":"ContainerDied","Data":"bef7de153dc8ea34e573c0cdbac5258398e4eda4bf0b59e92ba593d55dab3d9d"} Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.095120 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bef7de153dc8ea34e573c0cdbac5258398e4eda4bf0b59e92ba593d55dab3d9d" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.095016 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.205736 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg"] Dec 04 14:43:13 crc kubenswrapper[4848]: E1204 14:43:13.206346 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e25768-61f4-4749-9a5c-006da09aff30" containerName="registry-server" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.206367 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e25768-61f4-4749-9a5c-006da09aff30" containerName="registry-server" Dec 04 14:43:13 crc kubenswrapper[4848]: E1204 14:43:13.206388 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e25768-61f4-4749-9a5c-006da09aff30" containerName="extract-utilities" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.206397 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e25768-61f4-4749-9a5c-006da09aff30" containerName="extract-utilities" Dec 04 14:43:13 crc kubenswrapper[4848]: E1204 14:43:13.206430 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e25768-61f4-4749-9a5c-006da09aff30" containerName="extract-content" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.206438 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e25768-61f4-4749-9a5c-006da09aff30" containerName="extract-content" Dec 04 14:43:13 crc kubenswrapper[4848]: E1204 14:43:13.206484 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e112a8b-6965-4e32-bb39-4ff08519bdee" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.206495 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e112a8b-6965-4e32-bb39-4ff08519bdee" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.206793 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e25768-61f4-4749-9a5c-006da09aff30" containerName="registry-server" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.206828 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e112a8b-6965-4e32-bb39-4ff08519bdee" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.207675 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.210586 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.210798 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4cc42" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.211924 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.212234 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.212399 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.218053 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg"] Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.223607 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.223780 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.223873 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.223918 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzxj2\" (UniqueName: \"kubernetes.io/projected/d0800b88-1124-4b08-a53c-ba88de0acc98-kube-api-access-tzxj2\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.224230 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.327107 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.327394 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.327495 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzxj2\" (UniqueName: \"kubernetes.io/projected/d0800b88-1124-4b08-a53c-ba88de0acc98-kube-api-access-tzxj2\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.327651 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.327887 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.331611 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.331995 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.332115 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.336749 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.345511 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzxj2\" (UniqueName: \"kubernetes.io/projected/d0800b88-1124-4b08-a53c-ba88de0acc98-kube-api-access-tzxj2\") pod \"logging-edpm-deployment-openstack-edpm-ipam-9zcjg\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:13 crc kubenswrapper[4848]: I1204 14:43:13.527140 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:14 crc kubenswrapper[4848]: I1204 14:43:14.059027 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg"] Dec 04 14:43:14 crc kubenswrapper[4848]: I1204 14:43:14.108199 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" event={"ID":"d0800b88-1124-4b08-a53c-ba88de0acc98","Type":"ContainerStarted","Data":"96fc43e1c82104a1580ddb070322ee444d0b68d3ba61574a370af4e2626611b2"} Dec 04 14:43:15 crc kubenswrapper[4848]: I1204 14:43:15.122114 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" event={"ID":"d0800b88-1124-4b08-a53c-ba88de0acc98","Type":"ContainerStarted","Data":"3a809cde13ceef45510045cd7f5adc2c511b26997c41c4cdf1af0ebd4af116c3"} Dec 04 14:43:15 crc kubenswrapper[4848]: I1204 14:43:15.148903 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" podStartSLOduration=1.7259358420000002 podStartE2EDuration="2.14888349s" podCreationTimestamp="2025-12-04 14:43:13 +0000 UTC" firstStartedPulling="2025-12-04 14:43:14.061816625 +0000 UTC m=+3298.004313153" lastFinishedPulling="2025-12-04 14:43:14.484764283 +0000 UTC m=+3298.427260801" observedRunningTime="2025-12-04 14:43:15.138572498 +0000 UTC m=+3299.081069036" watchObservedRunningTime="2025-12-04 14:43:15.14888349 +0000 UTC m=+3299.091380018" Dec 04 14:43:24 crc kubenswrapper[4848]: I1204 14:43:24.393396 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:43:24 crc kubenswrapper[4848]: E1204 14:43:24.394203 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:43:29 crc kubenswrapper[4848]: I1204 14:43:29.278010 4848 generic.go:334] "Generic (PLEG): container finished" podID="d0800b88-1124-4b08-a53c-ba88de0acc98" containerID="3a809cde13ceef45510045cd7f5adc2c511b26997c41c4cdf1af0ebd4af116c3" exitCode=0 Dec 04 14:43:29 crc kubenswrapper[4848]: I1204 14:43:29.278056 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" event={"ID":"d0800b88-1124-4b08-a53c-ba88de0acc98","Type":"ContainerDied","Data":"3a809cde13ceef45510045cd7f5adc2c511b26997c41c4cdf1af0ebd4af116c3"} Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.721917 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.767044 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-inventory\") pod \"d0800b88-1124-4b08-a53c-ba88de0acc98\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.767167 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-0\") pod \"d0800b88-1124-4b08-a53c-ba88de0acc98\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.767202 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzxj2\" (UniqueName: \"kubernetes.io/projected/d0800b88-1124-4b08-a53c-ba88de0acc98-kube-api-access-tzxj2\") pod \"d0800b88-1124-4b08-a53c-ba88de0acc98\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.767414 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-ssh-key\") pod \"d0800b88-1124-4b08-a53c-ba88de0acc98\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.767563 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-1\") pod \"d0800b88-1124-4b08-a53c-ba88de0acc98\" (UID: \"d0800b88-1124-4b08-a53c-ba88de0acc98\") " Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.774110 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0800b88-1124-4b08-a53c-ba88de0acc98-kube-api-access-tzxj2" (OuterVolumeSpecName: "kube-api-access-tzxj2") pod "d0800b88-1124-4b08-a53c-ba88de0acc98" (UID: "d0800b88-1124-4b08-a53c-ba88de0acc98"). InnerVolumeSpecName "kube-api-access-tzxj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.803368 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-inventory" (OuterVolumeSpecName: "inventory") pod "d0800b88-1124-4b08-a53c-ba88de0acc98" (UID: "d0800b88-1124-4b08-a53c-ba88de0acc98"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.804005 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "d0800b88-1124-4b08-a53c-ba88de0acc98" (UID: "d0800b88-1124-4b08-a53c-ba88de0acc98"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.812111 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d0800b88-1124-4b08-a53c-ba88de0acc98" (UID: "d0800b88-1124-4b08-a53c-ba88de0acc98"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.812591 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "d0800b88-1124-4b08-a53c-ba88de0acc98" (UID: "d0800b88-1124-4b08-a53c-ba88de0acc98"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.869885 4848 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.869925 4848 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.869939 4848 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.869970 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzxj2\" (UniqueName: \"kubernetes.io/projected/d0800b88-1124-4b08-a53c-ba88de0acc98-kube-api-access-tzxj2\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:30 crc kubenswrapper[4848]: I1204 14:43:30.869982 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0800b88-1124-4b08-a53c-ba88de0acc98-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:43:31 crc kubenswrapper[4848]: I1204 14:43:31.337980 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" event={"ID":"d0800b88-1124-4b08-a53c-ba88de0acc98","Type":"ContainerDied","Data":"96fc43e1c82104a1580ddb070322ee444d0b68d3ba61574a370af4e2626611b2"} Dec 04 14:43:31 crc kubenswrapper[4848]: I1204 14:43:31.338227 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96fc43e1c82104a1580ddb070322ee444d0b68d3ba61574a370af4e2626611b2" Dec 04 14:43:31 crc kubenswrapper[4848]: I1204 14:43:31.338292 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-9zcjg" Dec 04 14:43:37 crc kubenswrapper[4848]: I1204 14:43:37.393965 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:43:37 crc kubenswrapper[4848]: E1204 14:43:37.394850 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:43:50 crc kubenswrapper[4848]: I1204 14:43:50.394019 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:43:51 crc kubenswrapper[4848]: I1204 14:43:51.651643 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"dfd7733c1fe0ac0c3b3ab55278f03f08d425de3ef89db3442f137b352ee5d4e4"} Dec 04 14:44:08 crc kubenswrapper[4848]: I1204 14:44:08.766492 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mwr2t"] Dec 04 14:44:08 crc kubenswrapper[4848]: E1204 14:44:08.767752 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0800b88-1124-4b08-a53c-ba88de0acc98" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 04 14:44:08 crc kubenswrapper[4848]: I1204 14:44:08.767769 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0800b88-1124-4b08-a53c-ba88de0acc98" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 04 14:44:08 crc kubenswrapper[4848]: I1204 14:44:08.768145 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0800b88-1124-4b08-a53c-ba88de0acc98" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 04 14:44:08 crc kubenswrapper[4848]: I1204 14:44:08.770482 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:08 crc kubenswrapper[4848]: I1204 14:44:08.779908 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwr2t"] Dec 04 14:44:08 crc kubenswrapper[4848]: I1204 14:44:08.914378 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjxv9\" (UniqueName: \"kubernetes.io/projected/7354b128-7a17-4a57-9492-4d0d691fafee-kube-api-access-rjxv9\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:08 crc kubenswrapper[4848]: I1204 14:44:08.914576 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-catalog-content\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:08 crc kubenswrapper[4848]: I1204 14:44:08.914674 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-utilities\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.017209 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-catalog-content\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.017468 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-utilities\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.017699 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjxv9\" (UniqueName: \"kubernetes.io/projected/7354b128-7a17-4a57-9492-4d0d691fafee-kube-api-access-rjxv9\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.017728 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-catalog-content\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.017928 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-utilities\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.038728 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjxv9\" (UniqueName: \"kubernetes.io/projected/7354b128-7a17-4a57-9492-4d0d691fafee-kube-api-access-rjxv9\") pod \"redhat-marketplace-mwr2t\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.108640 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.594901 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwr2t"] Dec 04 14:44:09 crc kubenswrapper[4848]: I1204 14:44:09.896566 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwr2t" event={"ID":"7354b128-7a17-4a57-9492-4d0d691fafee","Type":"ContainerStarted","Data":"e06238c5ac911290959329d20530ca2ba907002e9ba0df1a2e712365fd1eb312"} Dec 04 14:44:10 crc kubenswrapper[4848]: I1204 14:44:10.910179 4848 generic.go:334] "Generic (PLEG): container finished" podID="7354b128-7a17-4a57-9492-4d0d691fafee" containerID="f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e" exitCode=0 Dec 04 14:44:10 crc kubenswrapper[4848]: I1204 14:44:10.910512 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwr2t" event={"ID":"7354b128-7a17-4a57-9492-4d0d691fafee","Type":"ContainerDied","Data":"f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e"} Dec 04 14:44:10 crc kubenswrapper[4848]: I1204 14:44:10.913575 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:44:11 crc kubenswrapper[4848]: I1204 14:44:11.925517 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwr2t" event={"ID":"7354b128-7a17-4a57-9492-4d0d691fafee","Type":"ContainerStarted","Data":"0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24"} Dec 04 14:44:12 crc kubenswrapper[4848]: I1204 14:44:12.948880 4848 generic.go:334] "Generic (PLEG): container finished" podID="7354b128-7a17-4a57-9492-4d0d691fafee" containerID="0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24" exitCode=0 Dec 04 14:44:12 crc kubenswrapper[4848]: I1204 14:44:12.948942 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwr2t" event={"ID":"7354b128-7a17-4a57-9492-4d0d691fafee","Type":"ContainerDied","Data":"0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24"} Dec 04 14:44:13 crc kubenswrapper[4848]: I1204 14:44:13.965708 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwr2t" event={"ID":"7354b128-7a17-4a57-9492-4d0d691fafee","Type":"ContainerStarted","Data":"1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4"} Dec 04 14:44:14 crc kubenswrapper[4848]: I1204 14:44:14.003105 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mwr2t" podStartSLOduration=3.310225878 podStartE2EDuration="6.003087633s" podCreationTimestamp="2025-12-04 14:44:08 +0000 UTC" firstStartedPulling="2025-12-04 14:44:10.913291036 +0000 UTC m=+3354.855787564" lastFinishedPulling="2025-12-04 14:44:13.606152791 +0000 UTC m=+3357.548649319" observedRunningTime="2025-12-04 14:44:13.997777933 +0000 UTC m=+3357.940274461" watchObservedRunningTime="2025-12-04 14:44:14.003087633 +0000 UTC m=+3357.945584161" Dec 04 14:44:19 crc kubenswrapper[4848]: I1204 14:44:19.109428 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:19 crc kubenswrapper[4848]: I1204 14:44:19.109848 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:19 crc kubenswrapper[4848]: I1204 14:44:19.175593 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:20 crc kubenswrapper[4848]: I1204 14:44:20.100424 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:20 crc kubenswrapper[4848]: I1204 14:44:20.179761 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwr2t"] Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.062216 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mwr2t" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" containerName="registry-server" containerID="cri-o://1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4" gracePeriod=2 Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.704022 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.765061 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-catalog-content\") pod \"7354b128-7a17-4a57-9492-4d0d691fafee\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.765153 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-utilities\") pod \"7354b128-7a17-4a57-9492-4d0d691fafee\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.765262 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjxv9\" (UniqueName: \"kubernetes.io/projected/7354b128-7a17-4a57-9492-4d0d691fafee-kube-api-access-rjxv9\") pod \"7354b128-7a17-4a57-9492-4d0d691fafee\" (UID: \"7354b128-7a17-4a57-9492-4d0d691fafee\") " Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.767333 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-utilities" (OuterVolumeSpecName: "utilities") pod "7354b128-7a17-4a57-9492-4d0d691fafee" (UID: "7354b128-7a17-4a57-9492-4d0d691fafee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.771175 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7354b128-7a17-4a57-9492-4d0d691fafee-kube-api-access-rjxv9" (OuterVolumeSpecName: "kube-api-access-rjxv9") pod "7354b128-7a17-4a57-9492-4d0d691fafee" (UID: "7354b128-7a17-4a57-9492-4d0d691fafee"). InnerVolumeSpecName "kube-api-access-rjxv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.784223 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7354b128-7a17-4a57-9492-4d0d691fafee" (UID: "7354b128-7a17-4a57-9492-4d0d691fafee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.867695 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjxv9\" (UniqueName: \"kubernetes.io/projected/7354b128-7a17-4a57-9492-4d0d691fafee-kube-api-access-rjxv9\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.867725 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:22 crc kubenswrapper[4848]: I1204 14:44:22.867735 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7354b128-7a17-4a57-9492-4d0d691fafee-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.074635 4848 generic.go:334] "Generic (PLEG): container finished" podID="7354b128-7a17-4a57-9492-4d0d691fafee" containerID="1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4" exitCode=0 Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.074685 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwr2t" event={"ID":"7354b128-7a17-4a57-9492-4d0d691fafee","Type":"ContainerDied","Data":"1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4"} Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.074695 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwr2t" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.074723 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwr2t" event={"ID":"7354b128-7a17-4a57-9492-4d0d691fafee","Type":"ContainerDied","Data":"e06238c5ac911290959329d20530ca2ba907002e9ba0df1a2e712365fd1eb312"} Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.074745 4848 scope.go:117] "RemoveContainer" containerID="1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.109510 4848 scope.go:117] "RemoveContainer" containerID="0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.120690 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwr2t"] Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.131043 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwr2t"] Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.134211 4848 scope.go:117] "RemoveContainer" containerID="f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.191043 4848 scope.go:117] "RemoveContainer" containerID="1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4" Dec 04 14:44:23 crc kubenswrapper[4848]: E1204 14:44:23.191522 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4\": container with ID starting with 1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4 not found: ID does not exist" containerID="1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.191550 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4"} err="failed to get container status \"1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4\": rpc error: code = NotFound desc = could not find container \"1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4\": container with ID starting with 1b129f22922b5aff4e08467516264399ec601012711d11400a8d5c075da82bc4 not found: ID does not exist" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.191571 4848 scope.go:117] "RemoveContainer" containerID="0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24" Dec 04 14:44:23 crc kubenswrapper[4848]: E1204 14:44:23.191838 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24\": container with ID starting with 0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24 not found: ID does not exist" containerID="0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.191890 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24"} err="failed to get container status \"0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24\": rpc error: code = NotFound desc = could not find container \"0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24\": container with ID starting with 0a4f98b323a0574bf7dcd690dda239917be0e54a862542316a8005d30421cf24 not found: ID does not exist" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.191902 4848 scope.go:117] "RemoveContainer" containerID="f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e" Dec 04 14:44:23 crc kubenswrapper[4848]: E1204 14:44:23.192225 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e\": container with ID starting with f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e not found: ID does not exist" containerID="f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e" Dec 04 14:44:23 crc kubenswrapper[4848]: I1204 14:44:23.192249 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e"} err="failed to get container status \"f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e\": rpc error: code = NotFound desc = could not find container \"f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e\": container with ID starting with f212c50564a38fb89661c9d1b47f4f818050acd02e1ca00789b81036d0d0945e not found: ID does not exist" Dec 04 14:44:24 crc kubenswrapper[4848]: I1204 14:44:24.409694 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" path="/var/lib/kubelet/pods/7354b128-7a17-4a57-9492-4d0d691fafee/volumes" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.163567 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8"] Dec 04 14:45:00 crc kubenswrapper[4848]: E1204 14:45:00.164835 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" containerName="extract-content" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.164854 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" containerName="extract-content" Dec 04 14:45:00 crc kubenswrapper[4848]: E1204 14:45:00.164877 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" containerName="registry-server" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.164885 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" containerName="registry-server" Dec 04 14:45:00 crc kubenswrapper[4848]: E1204 14:45:00.164922 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" containerName="extract-utilities" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.164931 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" containerName="extract-utilities" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.165331 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="7354b128-7a17-4a57-9492-4d0d691fafee" containerName="registry-server" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.166535 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.176344 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.176753 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.184686 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8"] Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.283468 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ftbn\" (UniqueName: \"kubernetes.io/projected/92487436-fbca-4b93-b6ab-1fa9d5686522-kube-api-access-6ftbn\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.283893 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92487436-fbca-4b93-b6ab-1fa9d5686522-secret-volume\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.284139 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92487436-fbca-4b93-b6ab-1fa9d5686522-config-volume\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.386243 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ftbn\" (UniqueName: \"kubernetes.io/projected/92487436-fbca-4b93-b6ab-1fa9d5686522-kube-api-access-6ftbn\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.386770 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92487436-fbca-4b93-b6ab-1fa9d5686522-secret-volume\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.386856 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92487436-fbca-4b93-b6ab-1fa9d5686522-config-volume\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.387750 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92487436-fbca-4b93-b6ab-1fa9d5686522-config-volume\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.394762 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92487436-fbca-4b93-b6ab-1fa9d5686522-secret-volume\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.410796 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ftbn\" (UniqueName: \"kubernetes.io/projected/92487436-fbca-4b93-b6ab-1fa9d5686522-kube-api-access-6ftbn\") pod \"collect-profiles-29414325-nj2m8\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.505472 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:00 crc kubenswrapper[4848]: I1204 14:45:00.995768 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8"] Dec 04 14:45:01 crc kubenswrapper[4848]: I1204 14:45:01.509684 4848 generic.go:334] "Generic (PLEG): container finished" podID="92487436-fbca-4b93-b6ab-1fa9d5686522" containerID="ebf4b9428669e01c1301a64e72614a0485f61b5cd76dd2dca886f34eec536c76" exitCode=0 Dec 04 14:45:01 crc kubenswrapper[4848]: I1204 14:45:01.509771 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" event={"ID":"92487436-fbca-4b93-b6ab-1fa9d5686522","Type":"ContainerDied","Data":"ebf4b9428669e01c1301a64e72614a0485f61b5cd76dd2dca886f34eec536c76"} Dec 04 14:45:01 crc kubenswrapper[4848]: I1204 14:45:01.510001 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" event={"ID":"92487436-fbca-4b93-b6ab-1fa9d5686522","Type":"ContainerStarted","Data":"ae9076306ef4ef1b00ab4ec3cbdd78925faf90a6cf2b1d602f4d8d42a6ccfd7c"} Dec 04 14:45:02 crc kubenswrapper[4848]: I1204 14:45:02.892599 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:02 crc kubenswrapper[4848]: I1204 14:45:02.963157 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92487436-fbca-4b93-b6ab-1fa9d5686522-config-volume\") pod \"92487436-fbca-4b93-b6ab-1fa9d5686522\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " Dec 04 14:45:02 crc kubenswrapper[4848]: I1204 14:45:02.963306 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ftbn\" (UniqueName: \"kubernetes.io/projected/92487436-fbca-4b93-b6ab-1fa9d5686522-kube-api-access-6ftbn\") pod \"92487436-fbca-4b93-b6ab-1fa9d5686522\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " Dec 04 14:45:02 crc kubenswrapper[4848]: I1204 14:45:02.963457 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92487436-fbca-4b93-b6ab-1fa9d5686522-secret-volume\") pod \"92487436-fbca-4b93-b6ab-1fa9d5686522\" (UID: \"92487436-fbca-4b93-b6ab-1fa9d5686522\") " Dec 04 14:45:02 crc kubenswrapper[4848]: I1204 14:45:02.966643 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92487436-fbca-4b93-b6ab-1fa9d5686522-config-volume" (OuterVolumeSpecName: "config-volume") pod "92487436-fbca-4b93-b6ab-1fa9d5686522" (UID: "92487436-fbca-4b93-b6ab-1fa9d5686522"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:45:02 crc kubenswrapper[4848]: I1204 14:45:02.970657 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92487436-fbca-4b93-b6ab-1fa9d5686522-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "92487436-fbca-4b93-b6ab-1fa9d5686522" (UID: "92487436-fbca-4b93-b6ab-1fa9d5686522"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:45:02 crc kubenswrapper[4848]: I1204 14:45:02.989260 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92487436-fbca-4b93-b6ab-1fa9d5686522-kube-api-access-6ftbn" (OuterVolumeSpecName: "kube-api-access-6ftbn") pod "92487436-fbca-4b93-b6ab-1fa9d5686522" (UID: "92487436-fbca-4b93-b6ab-1fa9d5686522"). InnerVolumeSpecName "kube-api-access-6ftbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:45:03 crc kubenswrapper[4848]: I1204 14:45:03.074430 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92487436-fbca-4b93-b6ab-1fa9d5686522-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:03 crc kubenswrapper[4848]: I1204 14:45:03.074480 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ftbn\" (UniqueName: \"kubernetes.io/projected/92487436-fbca-4b93-b6ab-1fa9d5686522-kube-api-access-6ftbn\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:03 crc kubenswrapper[4848]: I1204 14:45:03.074497 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/92487436-fbca-4b93-b6ab-1fa9d5686522-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:03 crc kubenswrapper[4848]: I1204 14:45:03.532187 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" event={"ID":"92487436-fbca-4b93-b6ab-1fa9d5686522","Type":"ContainerDied","Data":"ae9076306ef4ef1b00ab4ec3cbdd78925faf90a6cf2b1d602f4d8d42a6ccfd7c"} Dec 04 14:45:03 crc kubenswrapper[4848]: I1204 14:45:03.532242 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae9076306ef4ef1b00ab4ec3cbdd78925faf90a6cf2b1d602f4d8d42a6ccfd7c" Dec 04 14:45:03 crc kubenswrapper[4848]: I1204 14:45:03.532280 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8" Dec 04 14:45:03 crc kubenswrapper[4848]: I1204 14:45:03.985379 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps"] Dec 04 14:45:03 crc kubenswrapper[4848]: I1204 14:45:03.999130 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-dgwps"] Dec 04 14:45:04 crc kubenswrapper[4848]: I1204 14:45:04.411999 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e790c562-9bc1-4b46-bea9-817210a2658a" path="/var/lib/kubelet/pods/e790c562-9bc1-4b46-bea9-817210a2658a/volumes" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.433877 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6xhb2"] Dec 04 14:45:18 crc kubenswrapper[4848]: E1204 14:45:18.435242 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92487436-fbca-4b93-b6ab-1fa9d5686522" containerName="collect-profiles" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.435265 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="92487436-fbca-4b93-b6ab-1fa9d5686522" containerName="collect-profiles" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.435731 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="92487436-fbca-4b93-b6ab-1fa9d5686522" containerName="collect-profiles" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.438418 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.453712 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6xhb2"] Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.554219 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-utilities\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.554488 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm6fn\" (UniqueName: \"kubernetes.io/projected/ca323b88-6f30-42fa-aa75-6cbb4b393d31-kube-api-access-zm6fn\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.554523 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-catalog-content\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.658280 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-utilities\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.658355 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm6fn\" (UniqueName: \"kubernetes.io/projected/ca323b88-6f30-42fa-aa75-6cbb4b393d31-kube-api-access-zm6fn\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.658393 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-catalog-content\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.659328 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-catalog-content\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.659535 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-utilities\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.682577 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm6fn\" (UniqueName: \"kubernetes.io/projected/ca323b88-6f30-42fa-aa75-6cbb4b393d31-kube-api-access-zm6fn\") pod \"redhat-operators-6xhb2\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:18 crc kubenswrapper[4848]: I1204 14:45:18.765282 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:19 crc kubenswrapper[4848]: I1204 14:45:19.267547 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6xhb2"] Dec 04 14:45:19 crc kubenswrapper[4848]: I1204 14:45:19.702387 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xhb2" event={"ID":"ca323b88-6f30-42fa-aa75-6cbb4b393d31","Type":"ContainerStarted","Data":"479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c"} Dec 04 14:45:19 crc kubenswrapper[4848]: I1204 14:45:19.702718 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xhb2" event={"ID":"ca323b88-6f30-42fa-aa75-6cbb4b393d31","Type":"ContainerStarted","Data":"890c57ca3e80db023dae456c57a115fc72ec97e060408e45c5b5f972e8676597"} Dec 04 14:45:20 crc kubenswrapper[4848]: I1204 14:45:20.714541 4848 generic.go:334] "Generic (PLEG): container finished" podID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerID="479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c" exitCode=0 Dec 04 14:45:20 crc kubenswrapper[4848]: I1204 14:45:20.714654 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xhb2" event={"ID":"ca323b88-6f30-42fa-aa75-6cbb4b393d31","Type":"ContainerDied","Data":"479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c"} Dec 04 14:45:21 crc kubenswrapper[4848]: I1204 14:45:21.729632 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xhb2" event={"ID":"ca323b88-6f30-42fa-aa75-6cbb4b393d31","Type":"ContainerStarted","Data":"7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c"} Dec 04 14:45:24 crc kubenswrapper[4848]: I1204 14:45:24.766502 4848 generic.go:334] "Generic (PLEG): container finished" podID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerID="7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c" exitCode=0 Dec 04 14:45:24 crc kubenswrapper[4848]: I1204 14:45:24.766570 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xhb2" event={"ID":"ca323b88-6f30-42fa-aa75-6cbb4b393d31","Type":"ContainerDied","Data":"7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c"} Dec 04 14:45:25 crc kubenswrapper[4848]: I1204 14:45:25.788158 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xhb2" event={"ID":"ca323b88-6f30-42fa-aa75-6cbb4b393d31","Type":"ContainerStarted","Data":"fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931"} Dec 04 14:45:25 crc kubenswrapper[4848]: I1204 14:45:25.810243 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6xhb2" podStartSLOduration=3.281739803 podStartE2EDuration="7.810226682s" podCreationTimestamp="2025-12-04 14:45:18 +0000 UTC" firstStartedPulling="2025-12-04 14:45:20.717019258 +0000 UTC m=+3424.659515786" lastFinishedPulling="2025-12-04 14:45:25.245506137 +0000 UTC m=+3429.188002665" observedRunningTime="2025-12-04 14:45:25.804169024 +0000 UTC m=+3429.746665552" watchObservedRunningTime="2025-12-04 14:45:25.810226682 +0000 UTC m=+3429.752723210" Dec 04 14:45:28 crc kubenswrapper[4848]: I1204 14:45:28.766311 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:28 crc kubenswrapper[4848]: I1204 14:45:28.766726 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:29 crc kubenswrapper[4848]: I1204 14:45:29.811711 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6xhb2" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="registry-server" probeResult="failure" output=< Dec 04 14:45:29 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 14:45:29 crc kubenswrapper[4848]: > Dec 04 14:45:38 crc kubenswrapper[4848]: I1204 14:45:38.100074 4848 scope.go:117] "RemoveContainer" containerID="1ef62e11ff1abc4703fc14131b12f21368c4cd30c0c3878948c0446505378348" Dec 04 14:45:38 crc kubenswrapper[4848]: I1204 14:45:38.816481 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:38 crc kubenswrapper[4848]: I1204 14:45:38.873005 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:39 crc kubenswrapper[4848]: I1204 14:45:39.053631 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6xhb2"] Dec 04 14:45:39 crc kubenswrapper[4848]: I1204 14:45:39.938504 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6xhb2" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="registry-server" containerID="cri-o://fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931" gracePeriod=2 Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.479486 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.633298 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-utilities\") pod \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.633369 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm6fn\" (UniqueName: \"kubernetes.io/projected/ca323b88-6f30-42fa-aa75-6cbb4b393d31-kube-api-access-zm6fn\") pod \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.633696 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-catalog-content\") pod \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\" (UID: \"ca323b88-6f30-42fa-aa75-6cbb4b393d31\") " Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.634269 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-utilities" (OuterVolumeSpecName: "utilities") pod "ca323b88-6f30-42fa-aa75-6cbb4b393d31" (UID: "ca323b88-6f30-42fa-aa75-6cbb4b393d31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.634569 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.640427 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca323b88-6f30-42fa-aa75-6cbb4b393d31-kube-api-access-zm6fn" (OuterVolumeSpecName: "kube-api-access-zm6fn") pod "ca323b88-6f30-42fa-aa75-6cbb4b393d31" (UID: "ca323b88-6f30-42fa-aa75-6cbb4b393d31"). InnerVolumeSpecName "kube-api-access-zm6fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.737128 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm6fn\" (UniqueName: \"kubernetes.io/projected/ca323b88-6f30-42fa-aa75-6cbb4b393d31-kube-api-access-zm6fn\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.739073 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca323b88-6f30-42fa-aa75-6cbb4b393d31" (UID: "ca323b88-6f30-42fa-aa75-6cbb4b393d31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.839709 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca323b88-6f30-42fa-aa75-6cbb4b393d31-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.953121 4848 generic.go:334] "Generic (PLEG): container finished" podID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerID="fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931" exitCode=0 Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.953168 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xhb2" event={"ID":"ca323b88-6f30-42fa-aa75-6cbb4b393d31","Type":"ContainerDied","Data":"fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931"} Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.953210 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xhb2" event={"ID":"ca323b88-6f30-42fa-aa75-6cbb4b393d31","Type":"ContainerDied","Data":"890c57ca3e80db023dae456c57a115fc72ec97e060408e45c5b5f972e8676597"} Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.953234 4848 scope.go:117] "RemoveContainer" containerID="fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.953232 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xhb2" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.984433 4848 scope.go:117] "RemoveContainer" containerID="7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c" Dec 04 14:45:40 crc kubenswrapper[4848]: I1204 14:45:40.994429 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6xhb2"] Dec 04 14:45:41 crc kubenswrapper[4848]: I1204 14:45:41.005663 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6xhb2"] Dec 04 14:45:41 crc kubenswrapper[4848]: I1204 14:45:41.015766 4848 scope.go:117] "RemoveContainer" containerID="479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c" Dec 04 14:45:41 crc kubenswrapper[4848]: I1204 14:45:41.068971 4848 scope.go:117] "RemoveContainer" containerID="fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931" Dec 04 14:45:41 crc kubenswrapper[4848]: E1204 14:45:41.069478 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931\": container with ID starting with fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931 not found: ID does not exist" containerID="fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931" Dec 04 14:45:41 crc kubenswrapper[4848]: I1204 14:45:41.069532 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931"} err="failed to get container status \"fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931\": rpc error: code = NotFound desc = could not find container \"fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931\": container with ID starting with fa8720f5a19201684ae5975c12497cb20e9b76eb1af6f03bf8cd4ca9377c4931 not found: ID does not exist" Dec 04 14:45:41 crc kubenswrapper[4848]: I1204 14:45:41.069579 4848 scope.go:117] "RemoveContainer" containerID="7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c" Dec 04 14:45:41 crc kubenswrapper[4848]: E1204 14:45:41.069937 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c\": container with ID starting with 7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c not found: ID does not exist" containerID="7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c" Dec 04 14:45:41 crc kubenswrapper[4848]: I1204 14:45:41.069992 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c"} err="failed to get container status \"7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c\": rpc error: code = NotFound desc = could not find container \"7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c\": container with ID starting with 7a855dd5bd0280f668fc4f217fa39c5816f4d29f6b3c1537ec38dd4eb67bf67c not found: ID does not exist" Dec 04 14:45:41 crc kubenswrapper[4848]: I1204 14:45:41.070019 4848 scope.go:117] "RemoveContainer" containerID="479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c" Dec 04 14:45:41 crc kubenswrapper[4848]: E1204 14:45:41.070296 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c\": container with ID starting with 479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c not found: ID does not exist" containerID="479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c" Dec 04 14:45:41 crc kubenswrapper[4848]: I1204 14:45:41.070316 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c"} err="failed to get container status \"479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c\": rpc error: code = NotFound desc = could not find container \"479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c\": container with ID starting with 479290f1f7cf53085ee534154ff7582adb21f30079c540200ebefc0663db677c not found: ID does not exist" Dec 04 14:45:42 crc kubenswrapper[4848]: I1204 14:45:42.408515 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" path="/var/lib/kubelet/pods/ca323b88-6f30-42fa-aa75-6cbb4b393d31/volumes" Dec 04 14:46:14 crc kubenswrapper[4848]: I1204 14:46:14.314377 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:46:14 crc kubenswrapper[4848]: I1204 14:46:14.315074 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:46:44 crc kubenswrapper[4848]: I1204 14:46:44.313816 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:46:44 crc kubenswrapper[4848]: I1204 14:46:44.314445 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.562909 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jhcqg"] Dec 04 14:46:56 crc kubenswrapper[4848]: E1204 14:46:56.564754 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="extract-content" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.564776 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="extract-content" Dec 04 14:46:56 crc kubenswrapper[4848]: E1204 14:46:56.564808 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="extract-utilities" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.564829 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="extract-utilities" Dec 04 14:46:56 crc kubenswrapper[4848]: E1204 14:46:56.564860 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="registry-server" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.564870 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="registry-server" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.565190 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca323b88-6f30-42fa-aa75-6cbb4b393d31" containerName="registry-server" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.568867 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.587365 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jhcqg"] Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.722023 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rvk7\" (UniqueName: \"kubernetes.io/projected/6a3b61e8-af67-418f-ae8c-a686bf5fce75-kube-api-access-5rvk7\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.722523 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-catalog-content\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.722753 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-utilities\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.824570 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-catalog-content\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.825028 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-utilities\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.825168 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-catalog-content\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.825346 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rvk7\" (UniqueName: \"kubernetes.io/projected/6a3b61e8-af67-418f-ae8c-a686bf5fce75-kube-api-access-5rvk7\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.825582 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-utilities\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.846251 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rvk7\" (UniqueName: \"kubernetes.io/projected/6a3b61e8-af67-418f-ae8c-a686bf5fce75-kube-api-access-5rvk7\") pod \"certified-operators-jhcqg\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:56 crc kubenswrapper[4848]: I1204 14:46:56.891828 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:46:57 crc kubenswrapper[4848]: I1204 14:46:57.521533 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jhcqg"] Dec 04 14:46:57 crc kubenswrapper[4848]: W1204 14:46:57.522400 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a3b61e8_af67_418f_ae8c_a686bf5fce75.slice/crio-d6349fad8da027ea5dd26e7d2356cbe2601c880e6a2e2bcf6256dfbdc506c48d WatchSource:0}: Error finding container d6349fad8da027ea5dd26e7d2356cbe2601c880e6a2e2bcf6256dfbdc506c48d: Status 404 returned error can't find the container with id d6349fad8da027ea5dd26e7d2356cbe2601c880e6a2e2bcf6256dfbdc506c48d Dec 04 14:46:57 crc kubenswrapper[4848]: I1204 14:46:57.812295 4848 generic.go:334] "Generic (PLEG): container finished" podID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerID="a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9" exitCode=0 Dec 04 14:46:57 crc kubenswrapper[4848]: I1204 14:46:57.812850 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhcqg" event={"ID":"6a3b61e8-af67-418f-ae8c-a686bf5fce75","Type":"ContainerDied","Data":"a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9"} Dec 04 14:46:57 crc kubenswrapper[4848]: I1204 14:46:57.815437 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhcqg" event={"ID":"6a3b61e8-af67-418f-ae8c-a686bf5fce75","Type":"ContainerStarted","Data":"d6349fad8da027ea5dd26e7d2356cbe2601c880e6a2e2bcf6256dfbdc506c48d"} Dec 04 14:46:58 crc kubenswrapper[4848]: I1204 14:46:58.829509 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhcqg" event={"ID":"6a3b61e8-af67-418f-ae8c-a686bf5fce75","Type":"ContainerStarted","Data":"cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e"} Dec 04 14:47:00 crc kubenswrapper[4848]: I1204 14:47:00.855579 4848 generic.go:334] "Generic (PLEG): container finished" podID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerID="cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e" exitCode=0 Dec 04 14:47:00 crc kubenswrapper[4848]: I1204 14:47:00.855645 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhcqg" event={"ID":"6a3b61e8-af67-418f-ae8c-a686bf5fce75","Type":"ContainerDied","Data":"cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e"} Dec 04 14:47:01 crc kubenswrapper[4848]: I1204 14:47:01.870789 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhcqg" event={"ID":"6a3b61e8-af67-418f-ae8c-a686bf5fce75","Type":"ContainerStarted","Data":"1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5"} Dec 04 14:47:01 crc kubenswrapper[4848]: I1204 14:47:01.895871 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jhcqg" podStartSLOduration=2.421549557 podStartE2EDuration="5.895848532s" podCreationTimestamp="2025-12-04 14:46:56 +0000 UTC" firstStartedPulling="2025-12-04 14:46:57.817211616 +0000 UTC m=+3521.759708144" lastFinishedPulling="2025-12-04 14:47:01.291510591 +0000 UTC m=+3525.234007119" observedRunningTime="2025-12-04 14:47:01.891353322 +0000 UTC m=+3525.833849850" watchObservedRunningTime="2025-12-04 14:47:01.895848532 +0000 UTC m=+3525.838345060" Dec 04 14:47:06 crc kubenswrapper[4848]: I1204 14:47:06.892345 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:47:06 crc kubenswrapper[4848]: I1204 14:47:06.893057 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:47:06 crc kubenswrapper[4848]: I1204 14:47:06.952755 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:47:07 crc kubenswrapper[4848]: I1204 14:47:07.007925 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:47:07 crc kubenswrapper[4848]: I1204 14:47:07.550219 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jhcqg"] Dec 04 14:47:08 crc kubenswrapper[4848]: I1204 14:47:08.949538 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jhcqg" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerName="registry-server" containerID="cri-o://1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5" gracePeriod=2 Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.466343 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.539701 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rvk7\" (UniqueName: \"kubernetes.io/projected/6a3b61e8-af67-418f-ae8c-a686bf5fce75-kube-api-access-5rvk7\") pod \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.539990 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-catalog-content\") pod \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.540207 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-utilities\") pod \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\" (UID: \"6a3b61e8-af67-418f-ae8c-a686bf5fce75\") " Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.541340 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-utilities" (OuterVolumeSpecName: "utilities") pod "6a3b61e8-af67-418f-ae8c-a686bf5fce75" (UID: "6a3b61e8-af67-418f-ae8c-a686bf5fce75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.546525 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a3b61e8-af67-418f-ae8c-a686bf5fce75-kube-api-access-5rvk7" (OuterVolumeSpecName: "kube-api-access-5rvk7") pod "6a3b61e8-af67-418f-ae8c-a686bf5fce75" (UID: "6a3b61e8-af67-418f-ae8c-a686bf5fce75"). InnerVolumeSpecName "kube-api-access-5rvk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.588807 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a3b61e8-af67-418f-ae8c-a686bf5fce75" (UID: "6a3b61e8-af67-418f-ae8c-a686bf5fce75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.642171 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.642521 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rvk7\" (UniqueName: \"kubernetes.io/projected/6a3b61e8-af67-418f-ae8c-a686bf5fce75-kube-api-access-5rvk7\") on node \"crc\" DevicePath \"\"" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.642532 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a3b61e8-af67-418f-ae8c-a686bf5fce75-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.971992 4848 generic.go:334] "Generic (PLEG): container finished" podID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerID="1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5" exitCode=0 Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.972048 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhcqg" event={"ID":"6a3b61e8-af67-418f-ae8c-a686bf5fce75","Type":"ContainerDied","Data":"1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5"} Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.972080 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhcqg" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.972109 4848 scope.go:117] "RemoveContainer" containerID="1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5" Dec 04 14:47:09 crc kubenswrapper[4848]: I1204 14:47:09.972093 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhcqg" event={"ID":"6a3b61e8-af67-418f-ae8c-a686bf5fce75","Type":"ContainerDied","Data":"d6349fad8da027ea5dd26e7d2356cbe2601c880e6a2e2bcf6256dfbdc506c48d"} Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.006784 4848 scope.go:117] "RemoveContainer" containerID="cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e" Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.018627 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jhcqg"] Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.030443 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jhcqg"] Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.059046 4848 scope.go:117] "RemoveContainer" containerID="a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9" Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.102809 4848 scope.go:117] "RemoveContainer" containerID="1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5" Dec 04 14:47:10 crc kubenswrapper[4848]: E1204 14:47:10.103221 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5\": container with ID starting with 1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5 not found: ID does not exist" containerID="1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5" Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.103253 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5"} err="failed to get container status \"1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5\": rpc error: code = NotFound desc = could not find container \"1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5\": container with ID starting with 1b65aec693d49042c5e1ce3e75ade0d8e3ae98c79384ebccdf27169c71f1c2e5 not found: ID does not exist" Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.103280 4848 scope.go:117] "RemoveContainer" containerID="cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e" Dec 04 14:47:10 crc kubenswrapper[4848]: E1204 14:47:10.103518 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e\": container with ID starting with cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e not found: ID does not exist" containerID="cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e" Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.103544 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e"} err="failed to get container status \"cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e\": rpc error: code = NotFound desc = could not find container \"cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e\": container with ID starting with cc5e19b3ea8f9ea4c6e27de346e8fcb42ac75ede791e28cd82f05c09e38fbf0e not found: ID does not exist" Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.103560 4848 scope.go:117] "RemoveContainer" containerID="a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9" Dec 04 14:47:10 crc kubenswrapper[4848]: E1204 14:47:10.103823 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9\": container with ID starting with a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9 not found: ID does not exist" containerID="a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9" Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.103850 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9"} err="failed to get container status \"a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9\": rpc error: code = NotFound desc = could not find container \"a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9\": container with ID starting with a0b468c584fbf1d920a44ab1e84ce78da48172333d3a9d2c8b0417be3236cdb9 not found: ID does not exist" Dec 04 14:47:10 crc kubenswrapper[4848]: I1204 14:47:10.407697 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" path="/var/lib/kubelet/pods/6a3b61e8-af67-418f-ae8c-a686bf5fce75/volumes" Dec 04 14:47:14 crc kubenswrapper[4848]: I1204 14:47:14.314478 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:47:14 crc kubenswrapper[4848]: I1204 14:47:14.314797 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:47:14 crc kubenswrapper[4848]: I1204 14:47:14.314850 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:47:14 crc kubenswrapper[4848]: I1204 14:47:14.316490 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfd7733c1fe0ac0c3b3ab55278f03f08d425de3ef89db3442f137b352ee5d4e4"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:47:14 crc kubenswrapper[4848]: I1204 14:47:14.316574 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://dfd7733c1fe0ac0c3b3ab55278f03f08d425de3ef89db3442f137b352ee5d4e4" gracePeriod=600 Dec 04 14:47:15 crc kubenswrapper[4848]: I1204 14:47:15.028046 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="dfd7733c1fe0ac0c3b3ab55278f03f08d425de3ef89db3442f137b352ee5d4e4" exitCode=0 Dec 04 14:47:15 crc kubenswrapper[4848]: I1204 14:47:15.028122 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"dfd7733c1fe0ac0c3b3ab55278f03f08d425de3ef89db3442f137b352ee5d4e4"} Dec 04 14:47:15 crc kubenswrapper[4848]: I1204 14:47:15.028908 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460"} Dec 04 14:47:15 crc kubenswrapper[4848]: I1204 14:47:15.028986 4848 scope.go:117] "RemoveContainer" containerID="08d77d014dc0c95b258e912de296720ead2bf7ee19e26d99095abe7cb5d1575a" Dec 04 14:49:14 crc kubenswrapper[4848]: I1204 14:49:14.314452 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:49:14 crc kubenswrapper[4848]: I1204 14:49:14.315300 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:49:44 crc kubenswrapper[4848]: I1204 14:49:44.314601 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:49:44 crc kubenswrapper[4848]: I1204 14:49:44.316045 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:50:14 crc kubenswrapper[4848]: I1204 14:50:14.313968 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:50:14 crc kubenswrapper[4848]: I1204 14:50:14.314450 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:50:14 crc kubenswrapper[4848]: I1204 14:50:14.314493 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:50:14 crc kubenswrapper[4848]: I1204 14:50:14.315014 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:50:14 crc kubenswrapper[4848]: I1204 14:50:14.315081 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" gracePeriod=600 Dec 04 14:50:14 crc kubenswrapper[4848]: E1204 14:50:14.442915 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:50:15 crc kubenswrapper[4848]: I1204 14:50:15.009070 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" exitCode=0 Dec 04 14:50:15 crc kubenswrapper[4848]: I1204 14:50:15.009132 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460"} Dec 04 14:50:15 crc kubenswrapper[4848]: I1204 14:50:15.009630 4848 scope.go:117] "RemoveContainer" containerID="dfd7733c1fe0ac0c3b3ab55278f03f08d425de3ef89db3442f137b352ee5d4e4" Dec 04 14:50:15 crc kubenswrapper[4848]: I1204 14:50:15.010568 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:50:15 crc kubenswrapper[4848]: E1204 14:50:15.010842 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:50:25 crc kubenswrapper[4848]: I1204 14:50:25.394087 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:50:25 crc kubenswrapper[4848]: E1204 14:50:25.394892 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:50:26 crc kubenswrapper[4848]: E1204 14:50:26.872480 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 04 14:50:39 crc kubenswrapper[4848]: I1204 14:50:39.393941 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:50:39 crc kubenswrapper[4848]: E1204 14:50:39.397504 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:50:51 crc kubenswrapper[4848]: I1204 14:50:51.393293 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:50:51 crc kubenswrapper[4848]: E1204 14:50:51.394131 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:51:04 crc kubenswrapper[4848]: I1204 14:51:04.394489 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:51:04 crc kubenswrapper[4848]: E1204 14:51:04.396034 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:51:15 crc kubenswrapper[4848]: I1204 14:51:15.393921 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:51:15 crc kubenswrapper[4848]: E1204 14:51:15.396465 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:51:29 crc kubenswrapper[4848]: I1204 14:51:29.394453 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:51:29 crc kubenswrapper[4848]: E1204 14:51:29.395991 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:51:44 crc kubenswrapper[4848]: I1204 14:51:44.395076 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:51:44 crc kubenswrapper[4848]: E1204 14:51:44.396376 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:51:55 crc kubenswrapper[4848]: I1204 14:51:55.394313 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:51:55 crc kubenswrapper[4848]: E1204 14:51:55.397086 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:52:10 crc kubenswrapper[4848]: I1204 14:52:10.394125 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:52:10 crc kubenswrapper[4848]: E1204 14:52:10.394810 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:52:25 crc kubenswrapper[4848]: I1204 14:52:25.394059 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:52:25 crc kubenswrapper[4848]: E1204 14:52:25.394892 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:52:37 crc kubenswrapper[4848]: I1204 14:52:37.393936 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:52:37 crc kubenswrapper[4848]: E1204 14:52:37.394731 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.324281 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8z6bj"] Dec 04 14:52:46 crc kubenswrapper[4848]: E1204 14:52:46.325356 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerName="registry-server" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.325372 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerName="registry-server" Dec 04 14:52:46 crc kubenswrapper[4848]: E1204 14:52:46.325402 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerName="extract-utilities" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.325409 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerName="extract-utilities" Dec 04 14:52:46 crc kubenswrapper[4848]: E1204 14:52:46.325423 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerName="extract-content" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.325429 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerName="extract-content" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.325677 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a3b61e8-af67-418f-ae8c-a686bf5fce75" containerName="registry-server" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.327423 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.339498 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8z6bj"] Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.375360 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-catalog-content\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.375835 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-utilities\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.376040 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbgt8\" (UniqueName: \"kubernetes.io/projected/edce5312-7e72-491c-9f2e-07f6c6ce46e2-kube-api-access-kbgt8\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.479669 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-utilities\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.479930 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbgt8\" (UniqueName: \"kubernetes.io/projected/edce5312-7e72-491c-9f2e-07f6c6ce46e2-kube-api-access-kbgt8\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.480178 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-catalog-content\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.482895 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-utilities\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.483863 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-catalog-content\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.507268 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbgt8\" (UniqueName: \"kubernetes.io/projected/edce5312-7e72-491c-9f2e-07f6c6ce46e2-kube-api-access-kbgt8\") pod \"community-operators-8z6bj\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:46 crc kubenswrapper[4848]: I1204 14:52:46.652162 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:47 crc kubenswrapper[4848]: I1204 14:52:47.337922 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8z6bj"] Dec 04 14:52:47 crc kubenswrapper[4848]: I1204 14:52:47.667913 4848 generic.go:334] "Generic (PLEG): container finished" podID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerID="955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b" exitCode=0 Dec 04 14:52:47 crc kubenswrapper[4848]: I1204 14:52:47.668288 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8z6bj" event={"ID":"edce5312-7e72-491c-9f2e-07f6c6ce46e2","Type":"ContainerDied","Data":"955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b"} Dec 04 14:52:47 crc kubenswrapper[4848]: I1204 14:52:47.668392 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8z6bj" event={"ID":"edce5312-7e72-491c-9f2e-07f6c6ce46e2","Type":"ContainerStarted","Data":"4854baa6e2480cf8c2893958e13e2659dca35fbed1b7f6d3c58b99237c71c141"} Dec 04 14:52:47 crc kubenswrapper[4848]: I1204 14:52:47.670979 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:52:48 crc kubenswrapper[4848]: I1204 14:52:48.679246 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8z6bj" event={"ID":"edce5312-7e72-491c-9f2e-07f6c6ce46e2","Type":"ContainerStarted","Data":"9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9"} Dec 04 14:52:49 crc kubenswrapper[4848]: I1204 14:52:49.393159 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:52:49 crc kubenswrapper[4848]: E1204 14:52:49.393802 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:52:51 crc kubenswrapper[4848]: I1204 14:52:51.726712 4848 generic.go:334] "Generic (PLEG): container finished" podID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerID="9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9" exitCode=0 Dec 04 14:52:51 crc kubenswrapper[4848]: I1204 14:52:51.726806 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8z6bj" event={"ID":"edce5312-7e72-491c-9f2e-07f6c6ce46e2","Type":"ContainerDied","Data":"9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9"} Dec 04 14:52:52 crc kubenswrapper[4848]: I1204 14:52:52.740743 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8z6bj" event={"ID":"edce5312-7e72-491c-9f2e-07f6c6ce46e2","Type":"ContainerStarted","Data":"fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746"} Dec 04 14:52:52 crc kubenswrapper[4848]: I1204 14:52:52.762973 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8z6bj" podStartSLOduration=2.296493034 podStartE2EDuration="6.762933664s" podCreationTimestamp="2025-12-04 14:52:46 +0000 UTC" firstStartedPulling="2025-12-04 14:52:47.670677349 +0000 UTC m=+3871.613173877" lastFinishedPulling="2025-12-04 14:52:52.137117979 +0000 UTC m=+3876.079614507" observedRunningTime="2025-12-04 14:52:52.758151167 +0000 UTC m=+3876.700647705" watchObservedRunningTime="2025-12-04 14:52:52.762933664 +0000 UTC m=+3876.705430202" Dec 04 14:52:56 crc kubenswrapper[4848]: I1204 14:52:56.654705 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:56 crc kubenswrapper[4848]: I1204 14:52:56.655055 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:52:56 crc kubenswrapper[4848]: I1204 14:52:56.722247 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:53:01 crc kubenswrapper[4848]: I1204 14:53:01.393456 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:53:01 crc kubenswrapper[4848]: E1204 14:53:01.394243 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:53:06 crc kubenswrapper[4848]: I1204 14:53:06.701733 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:53:06 crc kubenswrapper[4848]: I1204 14:53:06.755058 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8z6bj"] Dec 04 14:53:06 crc kubenswrapper[4848]: I1204 14:53:06.908225 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8z6bj" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerName="registry-server" containerID="cri-o://fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746" gracePeriod=2 Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.483319 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.582651 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbgt8\" (UniqueName: \"kubernetes.io/projected/edce5312-7e72-491c-9f2e-07f6c6ce46e2-kube-api-access-kbgt8\") pod \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.582891 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-catalog-content\") pod \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.583020 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-utilities\") pod \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\" (UID: \"edce5312-7e72-491c-9f2e-07f6c6ce46e2\") " Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.584213 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-utilities" (OuterVolumeSpecName: "utilities") pod "edce5312-7e72-491c-9f2e-07f6c6ce46e2" (UID: "edce5312-7e72-491c-9f2e-07f6c6ce46e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.592444 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edce5312-7e72-491c-9f2e-07f6c6ce46e2-kube-api-access-kbgt8" (OuterVolumeSpecName: "kube-api-access-kbgt8") pod "edce5312-7e72-491c-9f2e-07f6c6ce46e2" (UID: "edce5312-7e72-491c-9f2e-07f6c6ce46e2"). InnerVolumeSpecName "kube-api-access-kbgt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.637914 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edce5312-7e72-491c-9f2e-07f6c6ce46e2" (UID: "edce5312-7e72-491c-9f2e-07f6c6ce46e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.686108 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbgt8\" (UniqueName: \"kubernetes.io/projected/edce5312-7e72-491c-9f2e-07f6c6ce46e2-kube-api-access-kbgt8\") on node \"crc\" DevicePath \"\"" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.686151 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.686162 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edce5312-7e72-491c-9f2e-07f6c6ce46e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.920807 4848 generic.go:334] "Generic (PLEG): container finished" podID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerID="fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746" exitCode=0 Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.920860 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8z6bj" event={"ID":"edce5312-7e72-491c-9f2e-07f6c6ce46e2","Type":"ContainerDied","Data":"fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746"} Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.920865 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8z6bj" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.920899 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8z6bj" event={"ID":"edce5312-7e72-491c-9f2e-07f6c6ce46e2","Type":"ContainerDied","Data":"4854baa6e2480cf8c2893958e13e2659dca35fbed1b7f6d3c58b99237c71c141"} Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.920921 4848 scope.go:117] "RemoveContainer" containerID="fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.952311 4848 scope.go:117] "RemoveContainer" containerID="9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9" Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.954502 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8z6bj"] Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.971871 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8z6bj"] Dec 04 14:53:07 crc kubenswrapper[4848]: I1204 14:53:07.987938 4848 scope.go:117] "RemoveContainer" containerID="955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b" Dec 04 14:53:08 crc kubenswrapper[4848]: I1204 14:53:08.038732 4848 scope.go:117] "RemoveContainer" containerID="fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746" Dec 04 14:53:08 crc kubenswrapper[4848]: E1204 14:53:08.039300 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746\": container with ID starting with fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746 not found: ID does not exist" containerID="fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746" Dec 04 14:53:08 crc kubenswrapper[4848]: I1204 14:53:08.039354 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746"} err="failed to get container status \"fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746\": rpc error: code = NotFound desc = could not find container \"fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746\": container with ID starting with fe85da324dcfd3f44c750001e458b45d5d45ee00acf76cbe8866beb58a430746 not found: ID does not exist" Dec 04 14:53:08 crc kubenswrapper[4848]: I1204 14:53:08.039380 4848 scope.go:117] "RemoveContainer" containerID="9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9" Dec 04 14:53:08 crc kubenswrapper[4848]: E1204 14:53:08.039691 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9\": container with ID starting with 9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9 not found: ID does not exist" containerID="9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9" Dec 04 14:53:08 crc kubenswrapper[4848]: I1204 14:53:08.039731 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9"} err="failed to get container status \"9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9\": rpc error: code = NotFound desc = could not find container \"9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9\": container with ID starting with 9c5d9ae5ba575f801bd57ec867db68d5a2bbe92628db9e3cf0d548cf90bce4d9 not found: ID does not exist" Dec 04 14:53:08 crc kubenswrapper[4848]: I1204 14:53:08.039760 4848 scope.go:117] "RemoveContainer" containerID="955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b" Dec 04 14:53:08 crc kubenswrapper[4848]: E1204 14:53:08.040136 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b\": container with ID starting with 955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b not found: ID does not exist" containerID="955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b" Dec 04 14:53:08 crc kubenswrapper[4848]: I1204 14:53:08.040173 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b"} err="failed to get container status \"955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b\": rpc error: code = NotFound desc = could not find container \"955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b\": container with ID starting with 955ffa9c254e533dd0fa8d1a11b4f711ef7328be14dca64f04ce063aef06cd9b not found: ID does not exist" Dec 04 14:53:08 crc kubenswrapper[4848]: I1204 14:53:08.406120 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" path="/var/lib/kubelet/pods/edce5312-7e72-491c-9f2e-07f6c6ce46e2/volumes" Dec 04 14:53:13 crc kubenswrapper[4848]: I1204 14:53:13.394096 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:53:13 crc kubenswrapper[4848]: E1204 14:53:13.394914 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:53:26 crc kubenswrapper[4848]: I1204 14:53:26.405129 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:53:26 crc kubenswrapper[4848]: E1204 14:53:26.406073 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:53:39 crc kubenswrapper[4848]: I1204 14:53:39.395702 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:53:39 crc kubenswrapper[4848]: E1204 14:53:39.396764 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:53:53 crc kubenswrapper[4848]: I1204 14:53:53.395777 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:53:53 crc kubenswrapper[4848]: E1204 14:53:53.398395 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:54:05 crc kubenswrapper[4848]: I1204 14:54:05.393382 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:54:05 crc kubenswrapper[4848]: E1204 14:54:05.394246 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:54:20 crc kubenswrapper[4848]: I1204 14:54:20.393868 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:54:20 crc kubenswrapper[4848]: E1204 14:54:20.394658 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:54:24 crc kubenswrapper[4848]: E1204 14:54:24.405280 4848 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.5:48382->38.102.83.5:35999: write tcp 38.102.83.5:48382->38.102.83.5:35999: write: broken pipe Dec 04 14:54:33 crc kubenswrapper[4848]: I1204 14:54:33.393770 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:54:33 crc kubenswrapper[4848]: E1204 14:54:33.395636 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:54:44 crc kubenswrapper[4848]: I1204 14:54:44.393649 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:54:44 crc kubenswrapper[4848]: E1204 14:54:44.394432 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:54:59 crc kubenswrapper[4848]: I1204 14:54:59.393375 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:54:59 crc kubenswrapper[4848]: E1204 14:54:59.394096 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:55:11 crc kubenswrapper[4848]: I1204 14:55:11.393490 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:55:11 crc kubenswrapper[4848]: E1204 14:55:11.394251 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 14:55:25 crc kubenswrapper[4848]: I1204 14:55:25.393876 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 14:55:26 crc kubenswrapper[4848]: I1204 14:55:26.593908 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"90d9363f3d9ccec1254aa660e03a9b536ba39c67d05276df745bd8de5503a569"} Dec 04 14:55:29 crc kubenswrapper[4848]: I1204 14:55:29.889750 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rshgl"] Dec 04 14:55:29 crc kubenswrapper[4848]: E1204 14:55:29.891862 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerName="extract-utilities" Dec 04 14:55:29 crc kubenswrapper[4848]: I1204 14:55:29.891886 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerName="extract-utilities" Dec 04 14:55:29 crc kubenswrapper[4848]: E1204 14:55:29.891979 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerName="registry-server" Dec 04 14:55:29 crc kubenswrapper[4848]: I1204 14:55:29.891989 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerName="registry-server" Dec 04 14:55:29 crc kubenswrapper[4848]: E1204 14:55:29.892015 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerName="extract-content" Dec 04 14:55:29 crc kubenswrapper[4848]: I1204 14:55:29.892024 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerName="extract-content" Dec 04 14:55:29 crc kubenswrapper[4848]: I1204 14:55:29.892373 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="edce5312-7e72-491c-9f2e-07f6c6ce46e2" containerName="registry-server" Dec 04 14:55:29 crc kubenswrapper[4848]: I1204 14:55:29.895726 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:29 crc kubenswrapper[4848]: I1204 14:55:29.901691 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rshgl"] Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.072080 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-utilities\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.072367 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2b24\" (UniqueName: \"kubernetes.io/projected/aa26cbd3-2425-4837-8f7d-4636aad8822c-kube-api-access-t2b24\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.072433 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-catalog-content\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.174049 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2b24\" (UniqueName: \"kubernetes.io/projected/aa26cbd3-2425-4837-8f7d-4636aad8822c-kube-api-access-t2b24\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.174094 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-catalog-content\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.174160 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-utilities\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.174570 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-utilities\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.174619 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-catalog-content\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.196409 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2b24\" (UniqueName: \"kubernetes.io/projected/aa26cbd3-2425-4837-8f7d-4636aad8822c-kube-api-access-t2b24\") pod \"redhat-marketplace-rshgl\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.224428 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:30 crc kubenswrapper[4848]: I1204 14:55:30.704718 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rshgl"] Dec 04 14:55:31 crc kubenswrapper[4848]: I1204 14:55:31.649204 4848 generic.go:334] "Generic (PLEG): container finished" podID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerID="ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173" exitCode=0 Dec 04 14:55:31 crc kubenswrapper[4848]: I1204 14:55:31.649315 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rshgl" event={"ID":"aa26cbd3-2425-4837-8f7d-4636aad8822c","Type":"ContainerDied","Data":"ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173"} Dec 04 14:55:31 crc kubenswrapper[4848]: I1204 14:55:31.649856 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rshgl" event={"ID":"aa26cbd3-2425-4837-8f7d-4636aad8822c","Type":"ContainerStarted","Data":"115fe44e99d60410e0624352e1b5cf40015b13a6abbf51a82ec78c0b7d02c4cf"} Dec 04 14:55:32 crc kubenswrapper[4848]: I1204 14:55:32.664162 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rshgl" event={"ID":"aa26cbd3-2425-4837-8f7d-4636aad8822c","Type":"ContainerStarted","Data":"f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1"} Dec 04 14:55:33 crc kubenswrapper[4848]: I1204 14:55:33.677895 4848 generic.go:334] "Generic (PLEG): container finished" podID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerID="f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1" exitCode=0 Dec 04 14:55:33 crc kubenswrapper[4848]: I1204 14:55:33.678056 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rshgl" event={"ID":"aa26cbd3-2425-4837-8f7d-4636aad8822c","Type":"ContainerDied","Data":"f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1"} Dec 04 14:55:34 crc kubenswrapper[4848]: I1204 14:55:34.690048 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rshgl" event={"ID":"aa26cbd3-2425-4837-8f7d-4636aad8822c","Type":"ContainerStarted","Data":"971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc"} Dec 04 14:55:34 crc kubenswrapper[4848]: I1204 14:55:34.712103 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rshgl" podStartSLOduration=3.258604469 podStartE2EDuration="5.712085335s" podCreationTimestamp="2025-12-04 14:55:29 +0000 UTC" firstStartedPulling="2025-12-04 14:55:31.651350376 +0000 UTC m=+4035.593846904" lastFinishedPulling="2025-12-04 14:55:34.104831222 +0000 UTC m=+4038.047327770" observedRunningTime="2025-12-04 14:55:34.705695189 +0000 UTC m=+4038.648191717" watchObservedRunningTime="2025-12-04 14:55:34.712085335 +0000 UTC m=+4038.654581863" Dec 04 14:55:40 crc kubenswrapper[4848]: I1204 14:55:40.224629 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:40 crc kubenswrapper[4848]: I1204 14:55:40.225740 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:40 crc kubenswrapper[4848]: I1204 14:55:40.281297 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:40 crc kubenswrapper[4848]: I1204 14:55:40.832359 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:40 crc kubenswrapper[4848]: I1204 14:55:40.901842 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rshgl"] Dec 04 14:55:42 crc kubenswrapper[4848]: I1204 14:55:42.780991 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rshgl" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerName="registry-server" containerID="cri-o://971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc" gracePeriod=2 Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.325355 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.521843 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-utilities\") pod \"aa26cbd3-2425-4837-8f7d-4636aad8822c\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.523191 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-utilities" (OuterVolumeSpecName: "utilities") pod "aa26cbd3-2425-4837-8f7d-4636aad8822c" (UID: "aa26cbd3-2425-4837-8f7d-4636aad8822c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.522055 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2b24\" (UniqueName: \"kubernetes.io/projected/aa26cbd3-2425-4837-8f7d-4636aad8822c-kube-api-access-t2b24\") pod \"aa26cbd3-2425-4837-8f7d-4636aad8822c\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.524037 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-catalog-content\") pod \"aa26cbd3-2425-4837-8f7d-4636aad8822c\" (UID: \"aa26cbd3-2425-4837-8f7d-4636aad8822c\") " Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.525573 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.530123 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa26cbd3-2425-4837-8f7d-4636aad8822c-kube-api-access-t2b24" (OuterVolumeSpecName: "kube-api-access-t2b24") pod "aa26cbd3-2425-4837-8f7d-4636aad8822c" (UID: "aa26cbd3-2425-4837-8f7d-4636aad8822c"). InnerVolumeSpecName "kube-api-access-t2b24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.555620 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa26cbd3-2425-4837-8f7d-4636aad8822c" (UID: "aa26cbd3-2425-4837-8f7d-4636aad8822c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.627831 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2b24\" (UniqueName: \"kubernetes.io/projected/aa26cbd3-2425-4837-8f7d-4636aad8822c-kube-api-access-t2b24\") on node \"crc\" DevicePath \"\"" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.627874 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa26cbd3-2425-4837-8f7d-4636aad8822c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.799711 4848 generic.go:334] "Generic (PLEG): container finished" podID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerID="971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc" exitCode=0 Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.799797 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rshgl" event={"ID":"aa26cbd3-2425-4837-8f7d-4636aad8822c","Type":"ContainerDied","Data":"971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc"} Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.799851 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rshgl" event={"ID":"aa26cbd3-2425-4837-8f7d-4636aad8822c","Type":"ContainerDied","Data":"115fe44e99d60410e0624352e1b5cf40015b13a6abbf51a82ec78c0b7d02c4cf"} Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.799870 4848 scope.go:117] "RemoveContainer" containerID="971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.801083 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rshgl" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.837461 4848 scope.go:117] "RemoveContainer" containerID="f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.858929 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rshgl"] Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.863365 4848 scope.go:117] "RemoveContainer" containerID="ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.869976 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rshgl"] Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.925863 4848 scope.go:117] "RemoveContainer" containerID="971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc" Dec 04 14:55:43 crc kubenswrapper[4848]: E1204 14:55:43.926389 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc\": container with ID starting with 971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc not found: ID does not exist" containerID="971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.926420 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc"} err="failed to get container status \"971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc\": rpc error: code = NotFound desc = could not find container \"971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc\": container with ID starting with 971a6b0a607f10a152e1c0f3350d99eaafcf27d0a78cfc46f9c1fd200ae7dedc not found: ID does not exist" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.926444 4848 scope.go:117] "RemoveContainer" containerID="f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1" Dec 04 14:55:43 crc kubenswrapper[4848]: E1204 14:55:43.926826 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1\": container with ID starting with f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1 not found: ID does not exist" containerID="f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.926873 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1"} err="failed to get container status \"f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1\": rpc error: code = NotFound desc = could not find container \"f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1\": container with ID starting with f0364eb7012c7563abf1d9722ccbe170a1800835bfaf01ae7e195ee43aae6af1 not found: ID does not exist" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.926902 4848 scope.go:117] "RemoveContainer" containerID="ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173" Dec 04 14:55:43 crc kubenswrapper[4848]: E1204 14:55:43.927308 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173\": container with ID starting with ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173 not found: ID does not exist" containerID="ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173" Dec 04 14:55:43 crc kubenswrapper[4848]: I1204 14:55:43.927355 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173"} err="failed to get container status \"ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173\": rpc error: code = NotFound desc = could not find container \"ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173\": container with ID starting with ca5b37a16405fa5d403c40aa7fcc749b59520c94fdfc635b5198d8d405078173 not found: ID does not exist" Dec 04 14:55:43 crc kubenswrapper[4848]: E1204 14:55:43.931333 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa26cbd3_2425_4837_8f7d_4636aad8822c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa26cbd3_2425_4837_8f7d_4636aad8822c.slice/crio-115fe44e99d60410e0624352e1b5cf40015b13a6abbf51a82ec78c0b7d02c4cf\": RecentStats: unable to find data in memory cache]" Dec 04 14:55:44 crc kubenswrapper[4848]: I1204 14:55:44.405426 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" path="/var/lib/kubelet/pods/aa26cbd3-2425-4837-8f7d-4636aad8822c/volumes" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.501918 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lq8mk"] Dec 04 14:57:30 crc kubenswrapper[4848]: E1204 14:57:30.503093 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerName="registry-server" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.503110 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerName="registry-server" Dec 04 14:57:30 crc kubenswrapper[4848]: E1204 14:57:30.503143 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerName="extract-utilities" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.503153 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerName="extract-utilities" Dec 04 14:57:30 crc kubenswrapper[4848]: E1204 14:57:30.503172 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerName="extract-content" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.503179 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerName="extract-content" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.503501 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa26cbd3-2425-4837-8f7d-4636aad8822c" containerName="registry-server" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.505896 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.525833 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lq8mk"] Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.681353 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1085de76-81b1-4779-97aa-cbffc93aed49-catalog-content\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.681658 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1085de76-81b1-4779-97aa-cbffc93aed49-utilities\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.681902 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fcjw\" (UniqueName: \"kubernetes.io/projected/1085de76-81b1-4779-97aa-cbffc93aed49-kube-api-access-7fcjw\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.783813 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fcjw\" (UniqueName: \"kubernetes.io/projected/1085de76-81b1-4779-97aa-cbffc93aed49-kube-api-access-7fcjw\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.783880 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1085de76-81b1-4779-97aa-cbffc93aed49-catalog-content\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.783975 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1085de76-81b1-4779-97aa-cbffc93aed49-utilities\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.784439 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1085de76-81b1-4779-97aa-cbffc93aed49-catalog-content\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.784478 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1085de76-81b1-4779-97aa-cbffc93aed49-utilities\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.808158 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fcjw\" (UniqueName: \"kubernetes.io/projected/1085de76-81b1-4779-97aa-cbffc93aed49-kube-api-access-7fcjw\") pod \"certified-operators-lq8mk\" (UID: \"1085de76-81b1-4779-97aa-cbffc93aed49\") " pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:30 crc kubenswrapper[4848]: I1204 14:57:30.834751 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:31 crc kubenswrapper[4848]: I1204 14:57:31.449096 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lq8mk"] Dec 04 14:57:31 crc kubenswrapper[4848]: I1204 14:57:31.956038 4848 generic.go:334] "Generic (PLEG): container finished" podID="1085de76-81b1-4779-97aa-cbffc93aed49" containerID="73d91600e708cc988617cc56712a876c58e54c0f6307c848f4576308463efc26" exitCode=0 Dec 04 14:57:31 crc kubenswrapper[4848]: I1204 14:57:31.956093 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lq8mk" event={"ID":"1085de76-81b1-4779-97aa-cbffc93aed49","Type":"ContainerDied","Data":"73d91600e708cc988617cc56712a876c58e54c0f6307c848f4576308463efc26"} Dec 04 14:57:31 crc kubenswrapper[4848]: I1204 14:57:31.956125 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lq8mk" event={"ID":"1085de76-81b1-4779-97aa-cbffc93aed49","Type":"ContainerStarted","Data":"9b06f4e578c2f463f7ecf05a99f01740e4272e147d1586b03b448a8cfa64ea4c"} Dec 04 14:57:40 crc kubenswrapper[4848]: I1204 14:57:40.043724 4848 generic.go:334] "Generic (PLEG): container finished" podID="1085de76-81b1-4779-97aa-cbffc93aed49" containerID="dcbc742bffcb196567a22598db89dd238f3305bca2d5681a94d9bc8f9eccdd19" exitCode=0 Dec 04 14:57:40 crc kubenswrapper[4848]: I1204 14:57:40.043770 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lq8mk" event={"ID":"1085de76-81b1-4779-97aa-cbffc93aed49","Type":"ContainerDied","Data":"dcbc742bffcb196567a22598db89dd238f3305bca2d5681a94d9bc8f9eccdd19"} Dec 04 14:57:41 crc kubenswrapper[4848]: I1204 14:57:41.059010 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lq8mk" event={"ID":"1085de76-81b1-4779-97aa-cbffc93aed49","Type":"ContainerStarted","Data":"645d62d907a5b472272acfa5d1e24de016afe2d791d5072751132cdf08c0cc67"} Dec 04 14:57:41 crc kubenswrapper[4848]: I1204 14:57:41.078427 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lq8mk" podStartSLOduration=2.613225812 podStartE2EDuration="11.078405757s" podCreationTimestamp="2025-12-04 14:57:30 +0000 UTC" firstStartedPulling="2025-12-04 14:57:31.958252265 +0000 UTC m=+4155.900748813" lastFinishedPulling="2025-12-04 14:57:40.42343223 +0000 UTC m=+4164.365928758" observedRunningTime="2025-12-04 14:57:41.075568128 +0000 UTC m=+4165.018064656" watchObservedRunningTime="2025-12-04 14:57:41.078405757 +0000 UTC m=+4165.020902295" Dec 04 14:57:44 crc kubenswrapper[4848]: I1204 14:57:44.314430 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:57:44 crc kubenswrapper[4848]: I1204 14:57:44.315037 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:57:50 crc kubenswrapper[4848]: I1204 14:57:50.834919 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:50 crc kubenswrapper[4848]: I1204 14:57:50.835493 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:50 crc kubenswrapper[4848]: I1204 14:57:50.892660 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:51 crc kubenswrapper[4848]: I1204 14:57:51.238303 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lq8mk" Dec 04 14:57:51 crc kubenswrapper[4848]: I1204 14:57:51.321297 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lq8mk"] Dec 04 14:57:51 crc kubenswrapper[4848]: I1204 14:57:51.382589 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rkwjn"] Dec 04 14:57:51 crc kubenswrapper[4848]: I1204 14:57:51.382825 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rkwjn" podUID="b8658e99-6b59-4722-90b2-7a18294722df" containerName="registry-server" containerID="cri-o://a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29" gracePeriod=2 Dec 04 14:57:51 crc kubenswrapper[4848]: I1204 14:57:51.976116 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.101735 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-catalog-content\") pod \"b8658e99-6b59-4722-90b2-7a18294722df\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.102401 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9psb\" (UniqueName: \"kubernetes.io/projected/b8658e99-6b59-4722-90b2-7a18294722df-kube-api-access-x9psb\") pod \"b8658e99-6b59-4722-90b2-7a18294722df\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.102554 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-utilities\") pod \"b8658e99-6b59-4722-90b2-7a18294722df\" (UID: \"b8658e99-6b59-4722-90b2-7a18294722df\") " Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.104369 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-utilities" (OuterVolumeSpecName: "utilities") pod "b8658e99-6b59-4722-90b2-7a18294722df" (UID: "b8658e99-6b59-4722-90b2-7a18294722df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.123975 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8658e99-6b59-4722-90b2-7a18294722df-kube-api-access-x9psb" (OuterVolumeSpecName: "kube-api-access-x9psb") pod "b8658e99-6b59-4722-90b2-7a18294722df" (UID: "b8658e99-6b59-4722-90b2-7a18294722df"). InnerVolumeSpecName "kube-api-access-x9psb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.196165 4848 generic.go:334] "Generic (PLEG): container finished" podID="b8658e99-6b59-4722-90b2-7a18294722df" containerID="a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29" exitCode=0 Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.197024 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkwjn" event={"ID":"b8658e99-6b59-4722-90b2-7a18294722df","Type":"ContainerDied","Data":"a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29"} Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.197042 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkwjn" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.197075 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkwjn" event={"ID":"b8658e99-6b59-4722-90b2-7a18294722df","Type":"ContainerDied","Data":"067ff03e5f70fefd05a5dfa31456333975ee4fd3401eb425a8271797f5c1625d"} Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.197102 4848 scope.go:117] "RemoveContainer" containerID="a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.205087 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.205115 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9psb\" (UniqueName: \"kubernetes.io/projected/b8658e99-6b59-4722-90b2-7a18294722df-kube-api-access-x9psb\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.242660 4848 scope.go:117] "RemoveContainer" containerID="cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.274810 4848 scope.go:117] "RemoveContainer" containerID="7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.302422 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8658e99-6b59-4722-90b2-7a18294722df" (UID: "b8658e99-6b59-4722-90b2-7a18294722df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.307283 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8658e99-6b59-4722-90b2-7a18294722df-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.321124 4848 scope.go:117] "RemoveContainer" containerID="a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29" Dec 04 14:57:52 crc kubenswrapper[4848]: E1204 14:57:52.321616 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29\": container with ID starting with a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29 not found: ID does not exist" containerID="a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.321728 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29"} err="failed to get container status \"a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29\": rpc error: code = NotFound desc = could not find container \"a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29\": container with ID starting with a529e90bcdf4ca008ffac9d1ea4470176973aee717e8d70f6ac7e35b4953eb29 not found: ID does not exist" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.321808 4848 scope.go:117] "RemoveContainer" containerID="cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8" Dec 04 14:57:52 crc kubenswrapper[4848]: E1204 14:57:52.322326 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8\": container with ID starting with cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8 not found: ID does not exist" containerID="cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.322367 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8"} err="failed to get container status \"cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8\": rpc error: code = NotFound desc = could not find container \"cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8\": container with ID starting with cf7b7bd1cd8f0d5e92ca4b33a5323a5ee2903d2a991ff1ed0b4b10b395647eb8 not found: ID does not exist" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.322413 4848 scope.go:117] "RemoveContainer" containerID="7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557" Dec 04 14:57:52 crc kubenswrapper[4848]: E1204 14:57:52.323899 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557\": container with ID starting with 7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557 not found: ID does not exist" containerID="7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.323939 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557"} err="failed to get container status \"7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557\": rpc error: code = NotFound desc = could not find container \"7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557\": container with ID starting with 7b08b6469829a9643128862ab7a8a68ae3486ebf9e1b76aec53fefd7af642557 not found: ID does not exist" Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.600533 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rkwjn"] Dec 04 14:57:52 crc kubenswrapper[4848]: I1204 14:57:52.614364 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rkwjn"] Dec 04 14:57:54 crc kubenswrapper[4848]: I1204 14:57:54.409067 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8658e99-6b59-4722-90b2-7a18294722df" path="/var/lib/kubelet/pods/b8658e99-6b59-4722-90b2-7a18294722df/volumes" Dec 04 14:58:14 crc kubenswrapper[4848]: I1204 14:58:14.314942 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:58:14 crc kubenswrapper[4848]: I1204 14:58:14.315499 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.314906 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.315430 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.315472 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.316301 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90d9363f3d9ccec1254aa660e03a9b536ba39c67d05276df745bd8de5503a569"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.316344 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://90d9363f3d9ccec1254aa660e03a9b536ba39c67d05276df745bd8de5503a569" gracePeriod=600 Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.777431 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="90d9363f3d9ccec1254aa660e03a9b536ba39c67d05276df745bd8de5503a569" exitCode=0 Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.777496 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"90d9363f3d9ccec1254aa660e03a9b536ba39c67d05276df745bd8de5503a569"} Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.777658 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404"} Dec 04 14:58:44 crc kubenswrapper[4848]: I1204 14:58:44.777677 4848 scope.go:117] "RemoveContainer" containerID="f837c1ae67a959998fd996b83df04b4717242d302eb2e4065351a1d7092fd460" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.180516 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr"] Dec 04 15:00:00 crc kubenswrapper[4848]: E1204 15:00:00.186663 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8658e99-6b59-4722-90b2-7a18294722df" containerName="extract-content" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.186707 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8658e99-6b59-4722-90b2-7a18294722df" containerName="extract-content" Dec 04 15:00:00 crc kubenswrapper[4848]: E1204 15:00:00.186754 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8658e99-6b59-4722-90b2-7a18294722df" containerName="extract-utilities" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.186762 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8658e99-6b59-4722-90b2-7a18294722df" containerName="extract-utilities" Dec 04 15:00:00 crc kubenswrapper[4848]: E1204 15:00:00.186788 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8658e99-6b59-4722-90b2-7a18294722df" containerName="registry-server" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.186794 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8658e99-6b59-4722-90b2-7a18294722df" containerName="registry-server" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.187155 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8658e99-6b59-4722-90b2-7a18294722df" containerName="registry-server" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.188152 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.190363 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.191708 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.194108 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr"] Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.312416 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/663026c2-88c0-4a7f-b92f-392fb9d7afe7-secret-volume\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.312828 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd2rl\" (UniqueName: \"kubernetes.io/projected/663026c2-88c0-4a7f-b92f-392fb9d7afe7-kube-api-access-nd2rl\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.313051 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/663026c2-88c0-4a7f-b92f-392fb9d7afe7-config-volume\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.415411 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/663026c2-88c0-4a7f-b92f-392fb9d7afe7-config-volume\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.415583 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/663026c2-88c0-4a7f-b92f-392fb9d7afe7-secret-volume\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.415637 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd2rl\" (UniqueName: \"kubernetes.io/projected/663026c2-88c0-4a7f-b92f-392fb9d7afe7-kube-api-access-nd2rl\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.416486 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/663026c2-88c0-4a7f-b92f-392fb9d7afe7-config-volume\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.421273 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/663026c2-88c0-4a7f-b92f-392fb9d7afe7-secret-volume\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.435894 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd2rl\" (UniqueName: \"kubernetes.io/projected/663026c2-88c0-4a7f-b92f-392fb9d7afe7-kube-api-access-nd2rl\") pod \"collect-profiles-29414340-5bpkr\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.523465 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:00 crc kubenswrapper[4848]: I1204 15:00:00.969547 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr"] Dec 04 15:00:00 crc kubenswrapper[4848]: W1204 15:00:00.982825 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod663026c2_88c0_4a7f_b92f_392fb9d7afe7.slice/crio-7bd8fa6edb9ca23f1740b6213b45d267d1a181bb239ddfdacc8195e6d86408b3 WatchSource:0}: Error finding container 7bd8fa6edb9ca23f1740b6213b45d267d1a181bb239ddfdacc8195e6d86408b3: Status 404 returned error can't find the container with id 7bd8fa6edb9ca23f1740b6213b45d267d1a181bb239ddfdacc8195e6d86408b3 Dec 04 15:00:01 crc kubenswrapper[4848]: I1204 15:00:01.719827 4848 generic.go:334] "Generic (PLEG): container finished" podID="663026c2-88c0-4a7f-b92f-392fb9d7afe7" containerID="87f332dca13d7bad2e654f1889b4af5b87fb7ef95904323e2c9513b42313887f" exitCode=0 Dec 04 15:00:01 crc kubenswrapper[4848]: I1204 15:00:01.719881 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" event={"ID":"663026c2-88c0-4a7f-b92f-392fb9d7afe7","Type":"ContainerDied","Data":"87f332dca13d7bad2e654f1889b4af5b87fb7ef95904323e2c9513b42313887f"} Dec 04 15:00:01 crc kubenswrapper[4848]: I1204 15:00:01.720999 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" event={"ID":"663026c2-88c0-4a7f-b92f-392fb9d7afe7","Type":"ContainerStarted","Data":"7bd8fa6edb9ca23f1740b6213b45d267d1a181bb239ddfdacc8195e6d86408b3"} Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.271966 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.398863 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/663026c2-88c0-4a7f-b92f-392fb9d7afe7-secret-volume\") pod \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.399269 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd2rl\" (UniqueName: \"kubernetes.io/projected/663026c2-88c0-4a7f-b92f-392fb9d7afe7-kube-api-access-nd2rl\") pod \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.399347 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/663026c2-88c0-4a7f-b92f-392fb9d7afe7-config-volume\") pod \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\" (UID: \"663026c2-88c0-4a7f-b92f-392fb9d7afe7\") " Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.400162 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/663026c2-88c0-4a7f-b92f-392fb9d7afe7-config-volume" (OuterVolumeSpecName: "config-volume") pod "663026c2-88c0-4a7f-b92f-392fb9d7afe7" (UID: "663026c2-88c0-4a7f-b92f-392fb9d7afe7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.404805 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/663026c2-88c0-4a7f-b92f-392fb9d7afe7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "663026c2-88c0-4a7f-b92f-392fb9d7afe7" (UID: "663026c2-88c0-4a7f-b92f-392fb9d7afe7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.405621 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/663026c2-88c0-4a7f-b92f-392fb9d7afe7-kube-api-access-nd2rl" (OuterVolumeSpecName: "kube-api-access-nd2rl") pod "663026c2-88c0-4a7f-b92f-392fb9d7afe7" (UID: "663026c2-88c0-4a7f-b92f-392fb9d7afe7"). InnerVolumeSpecName "kube-api-access-nd2rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.502081 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd2rl\" (UniqueName: \"kubernetes.io/projected/663026c2-88c0-4a7f-b92f-392fb9d7afe7-kube-api-access-nd2rl\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.502110 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/663026c2-88c0-4a7f-b92f-392fb9d7afe7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.502119 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/663026c2-88c0-4a7f-b92f-392fb9d7afe7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.745028 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" event={"ID":"663026c2-88c0-4a7f-b92f-392fb9d7afe7","Type":"ContainerDied","Data":"7bd8fa6edb9ca23f1740b6213b45d267d1a181bb239ddfdacc8195e6d86408b3"} Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.745323 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bd8fa6edb9ca23f1740b6213b45d267d1a181bb239ddfdacc8195e6d86408b3" Dec 04 15:00:03 crc kubenswrapper[4848]: I1204 15:00:03.745092 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr" Dec 04 15:00:04 crc kubenswrapper[4848]: I1204 15:00:04.349805 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4"] Dec 04 15:00:04 crc kubenswrapper[4848]: I1204 15:00:04.359680 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-8kqr4"] Dec 04 15:00:04 crc kubenswrapper[4848]: I1204 15:00:04.408433 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c" path="/var/lib/kubelet/pods/cb90a77b-84e6-4136-a1ff-8c4c3db2ce7c/volumes" Dec 04 15:00:38 crc kubenswrapper[4848]: I1204 15:00:38.552828 4848 scope.go:117] "RemoveContainer" containerID="7c2fadc0017816f05309698f024f113ac942635367ea9f5e829ffc9fee183022" Dec 04 15:00:44 crc kubenswrapper[4848]: I1204 15:00:44.314820 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:00:44 crc kubenswrapper[4848]: I1204 15:00:44.315495 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.191412 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29414341-72l4q"] Dec 04 15:01:00 crc kubenswrapper[4848]: E1204 15:01:00.192604 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="663026c2-88c0-4a7f-b92f-392fb9d7afe7" containerName="collect-profiles" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.192622 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="663026c2-88c0-4a7f-b92f-392fb9d7afe7" containerName="collect-profiles" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.192976 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="663026c2-88c0-4a7f-b92f-392fb9d7afe7" containerName="collect-profiles" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.193965 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.203775 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414341-72l4q"] Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.280327 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-combined-ca-bundle\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.280402 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8vp5\" (UniqueName: \"kubernetes.io/projected/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-kube-api-access-w8vp5\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.280636 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-config-data\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.280875 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-fernet-keys\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.383687 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-config-data\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.383768 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-fernet-keys\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.383812 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-combined-ca-bundle\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.383836 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8vp5\" (UniqueName: \"kubernetes.io/projected/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-kube-api-access-w8vp5\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.389798 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-config-data\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.389929 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-fernet-keys\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.391008 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-combined-ca-bundle\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.408634 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8vp5\" (UniqueName: \"kubernetes.io/projected/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-kube-api-access-w8vp5\") pod \"keystone-cron-29414341-72l4q\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.520011 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:00 crc kubenswrapper[4848]: I1204 15:01:00.995970 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414341-72l4q"] Dec 04 15:01:01 crc kubenswrapper[4848]: I1204 15:01:01.362922 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414341-72l4q" event={"ID":"9fe270e6-d360-4e3f-851e-3ab6b76b92a9","Type":"ContainerStarted","Data":"a5dc78e15bb195c0ab4bfdfb7a871a31956d1f077b3df2033f377ff7ec19254c"} Dec 04 15:01:01 crc kubenswrapper[4848]: I1204 15:01:01.363252 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414341-72l4q" event={"ID":"9fe270e6-d360-4e3f-851e-3ab6b76b92a9","Type":"ContainerStarted","Data":"59e032b8347bff2c6bbafc03f8d5c9e2e83ec01666c456283b6f93777c1c4f21"} Dec 04 15:01:01 crc kubenswrapper[4848]: I1204 15:01:01.378077 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29414341-72l4q" podStartSLOduration=1.378058839 podStartE2EDuration="1.378058839s" podCreationTimestamp="2025-12-04 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 15:01:01.3764677 +0000 UTC m=+4365.318964238" watchObservedRunningTime="2025-12-04 15:01:01.378058839 +0000 UTC m=+4365.320555367" Dec 04 15:01:05 crc kubenswrapper[4848]: I1204 15:01:05.412144 4848 generic.go:334] "Generic (PLEG): container finished" podID="9fe270e6-d360-4e3f-851e-3ab6b76b92a9" containerID="a5dc78e15bb195c0ab4bfdfb7a871a31956d1f077b3df2033f377ff7ec19254c" exitCode=0 Dec 04 15:01:05 crc kubenswrapper[4848]: I1204 15:01:05.412256 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414341-72l4q" event={"ID":"9fe270e6-d360-4e3f-851e-3ab6b76b92a9","Type":"ContainerDied","Data":"a5dc78e15bb195c0ab4bfdfb7a871a31956d1f077b3df2033f377ff7ec19254c"} Dec 04 15:01:06 crc kubenswrapper[4848]: I1204 15:01:06.900620 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.047539 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8vp5\" (UniqueName: \"kubernetes.io/projected/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-kube-api-access-w8vp5\") pod \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.047912 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-fernet-keys\") pod \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.048240 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-combined-ca-bundle\") pod \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.048344 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-config-data\") pod \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\" (UID: \"9fe270e6-d360-4e3f-851e-3ab6b76b92a9\") " Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.054702 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9fe270e6-d360-4e3f-851e-3ab6b76b92a9" (UID: "9fe270e6-d360-4e3f-851e-3ab6b76b92a9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.055114 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-kube-api-access-w8vp5" (OuterVolumeSpecName: "kube-api-access-w8vp5") pod "9fe270e6-d360-4e3f-851e-3ab6b76b92a9" (UID: "9fe270e6-d360-4e3f-851e-3ab6b76b92a9"). InnerVolumeSpecName "kube-api-access-w8vp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.085374 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fe270e6-d360-4e3f-851e-3ab6b76b92a9" (UID: "9fe270e6-d360-4e3f-851e-3ab6b76b92a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.119623 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-config-data" (OuterVolumeSpecName: "config-data") pod "9fe270e6-d360-4e3f-851e-3ab6b76b92a9" (UID: "9fe270e6-d360-4e3f-851e-3ab6b76b92a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.152708 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8vp5\" (UniqueName: \"kubernetes.io/projected/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-kube-api-access-w8vp5\") on node \"crc\" DevicePath \"\"" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.152736 4848 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.152746 4848 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.152757 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fe270e6-d360-4e3f-851e-3ab6b76b92a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.434714 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414341-72l4q" event={"ID":"9fe270e6-d360-4e3f-851e-3ab6b76b92a9","Type":"ContainerDied","Data":"59e032b8347bff2c6bbafc03f8d5c9e2e83ec01666c456283b6f93777c1c4f21"} Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.434772 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414341-72l4q" Dec 04 15:01:07 crc kubenswrapper[4848]: I1204 15:01:07.434782 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59e032b8347bff2c6bbafc03f8d5c9e2e83ec01666c456283b6f93777c1c4f21" Dec 04 15:01:14 crc kubenswrapper[4848]: I1204 15:01:14.314312 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:01:14 crc kubenswrapper[4848]: I1204 15:01:14.314935 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.314468 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.315047 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.315091 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.315998 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.316055 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" gracePeriod=600 Dec 04 15:01:44 crc kubenswrapper[4848]: E1204 15:01:44.436707 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.889842 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" exitCode=0 Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.890013 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404"} Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.890233 4848 scope.go:117] "RemoveContainer" containerID="90d9363f3d9ccec1254aa660e03a9b536ba39c67d05276df745bd8de5503a569" Dec 04 15:01:44 crc kubenswrapper[4848]: I1204 15:01:44.891117 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:01:44 crc kubenswrapper[4848]: E1204 15:01:44.891494 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.012472 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zfpmh"] Dec 04 15:01:55 crc kubenswrapper[4848]: E1204 15:01:55.018363 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe270e6-d360-4e3f-851e-3ab6b76b92a9" containerName="keystone-cron" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.018401 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe270e6-d360-4e3f-851e-3ab6b76b92a9" containerName="keystone-cron" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.025401 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fe270e6-d360-4e3f-851e-3ab6b76b92a9" containerName="keystone-cron" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.027370 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.037530 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfpmh"] Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.065378 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b179aa9a-174a-42ba-a529-9f2aab63c1e9-catalog-content\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.065449 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjk8w\" (UniqueName: \"kubernetes.io/projected/b179aa9a-174a-42ba-a529-9f2aab63c1e9-kube-api-access-zjk8w\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.065486 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b179aa9a-174a-42ba-a529-9f2aab63c1e9-utilities\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.168723 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b179aa9a-174a-42ba-a529-9f2aab63c1e9-catalog-content\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.168819 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjk8w\" (UniqueName: \"kubernetes.io/projected/b179aa9a-174a-42ba-a529-9f2aab63c1e9-kube-api-access-zjk8w\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.168872 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b179aa9a-174a-42ba-a529-9f2aab63c1e9-utilities\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.169578 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b179aa9a-174a-42ba-a529-9f2aab63c1e9-utilities\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.169589 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b179aa9a-174a-42ba-a529-9f2aab63c1e9-catalog-content\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.190916 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjk8w\" (UniqueName: \"kubernetes.io/projected/b179aa9a-174a-42ba-a529-9f2aab63c1e9-kube-api-access-zjk8w\") pod \"redhat-operators-zfpmh\" (UID: \"b179aa9a-174a-42ba-a529-9f2aab63c1e9\") " pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.366023 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:01:55 crc kubenswrapper[4848]: I1204 15:01:55.919868 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfpmh"] Dec 04 15:01:56 crc kubenswrapper[4848]: I1204 15:01:56.045146 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfpmh" event={"ID":"b179aa9a-174a-42ba-a529-9f2aab63c1e9","Type":"ContainerStarted","Data":"fa49dc7dcb36b14d1a272d0096b88de87eb7fef33bbbe34262224fcf7602fa2c"} Dec 04 15:01:57 crc kubenswrapper[4848]: I1204 15:01:57.057487 4848 generic.go:334] "Generic (PLEG): container finished" podID="b179aa9a-174a-42ba-a529-9f2aab63c1e9" containerID="d698e606fb1d8f5fa11e723e8b4170f205028724b3e888e80a0abee723e44927" exitCode=0 Dec 04 15:01:57 crc kubenswrapper[4848]: I1204 15:01:57.057584 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfpmh" event={"ID":"b179aa9a-174a-42ba-a529-9f2aab63c1e9","Type":"ContainerDied","Data":"d698e606fb1d8f5fa11e723e8b4170f205028724b3e888e80a0abee723e44927"} Dec 04 15:01:57 crc kubenswrapper[4848]: I1204 15:01:57.059672 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:01:58 crc kubenswrapper[4848]: I1204 15:01:58.393717 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:01:58 crc kubenswrapper[4848]: E1204 15:01:58.394234 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:02:07 crc kubenswrapper[4848]: I1204 15:02:07.203834 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfpmh" event={"ID":"b179aa9a-174a-42ba-a529-9f2aab63c1e9","Type":"ContainerStarted","Data":"dd59be74bdaa263998a73823a4b837779d7f1cf41f045db1c6ee6b435349eb27"} Dec 04 15:02:12 crc kubenswrapper[4848]: I1204 15:02:12.281316 4848 generic.go:334] "Generic (PLEG): container finished" podID="b179aa9a-174a-42ba-a529-9f2aab63c1e9" containerID="dd59be74bdaa263998a73823a4b837779d7f1cf41f045db1c6ee6b435349eb27" exitCode=0 Dec 04 15:02:12 crc kubenswrapper[4848]: I1204 15:02:12.281418 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfpmh" event={"ID":"b179aa9a-174a-42ba-a529-9f2aab63c1e9","Type":"ContainerDied","Data":"dd59be74bdaa263998a73823a4b837779d7f1cf41f045db1c6ee6b435349eb27"} Dec 04 15:02:13 crc kubenswrapper[4848]: I1204 15:02:13.304789 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfpmh" event={"ID":"b179aa9a-174a-42ba-a529-9f2aab63c1e9","Type":"ContainerStarted","Data":"72de0203ba07c395526c64b8d8e8523325672b35dfb561c2599121c12adf54e0"} Dec 04 15:02:13 crc kubenswrapper[4848]: I1204 15:02:13.338848 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zfpmh" podStartSLOduration=3.6429160659999997 podStartE2EDuration="19.338820721s" podCreationTimestamp="2025-12-04 15:01:54 +0000 UTC" firstStartedPulling="2025-12-04 15:01:57.059451727 +0000 UTC m=+4421.001948255" lastFinishedPulling="2025-12-04 15:02:12.755356382 +0000 UTC m=+4436.697852910" observedRunningTime="2025-12-04 15:02:13.327912413 +0000 UTC m=+4437.270408941" watchObservedRunningTime="2025-12-04 15:02:13.338820721 +0000 UTC m=+4437.281317249" Dec 04 15:02:13 crc kubenswrapper[4848]: I1204 15:02:13.394535 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:02:13 crc kubenswrapper[4848]: E1204 15:02:13.394922 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:02:15 crc kubenswrapper[4848]: I1204 15:02:15.366647 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:02:15 crc kubenswrapper[4848]: I1204 15:02:15.367009 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:02:16 crc kubenswrapper[4848]: I1204 15:02:16.800600 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zfpmh" podUID="b179aa9a-174a-42ba-a529-9f2aab63c1e9" containerName="registry-server" probeResult="failure" output=< Dec 04 15:02:16 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:02:16 crc kubenswrapper[4848]: > Dec 04 15:02:25 crc kubenswrapper[4848]: I1204 15:02:25.422685 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:02:25 crc kubenswrapper[4848]: I1204 15:02:25.478103 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zfpmh" Dec 04 15:02:26 crc kubenswrapper[4848]: I1204 15:02:26.043011 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfpmh"] Dec 04 15:02:26 crc kubenswrapper[4848]: I1204 15:02:26.220134 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q2ntw"] Dec 04 15:02:26 crc kubenswrapper[4848]: I1204 15:02:26.220429 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q2ntw" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="registry-server" containerID="cri-o://9c6726243aa9ce482f6bef9410b748d2b9445fc325591a6cf970e2f1dbc83b9b" gracePeriod=2 Dec 04 15:02:26 crc kubenswrapper[4848]: I1204 15:02:26.401810 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:02:26 crc kubenswrapper[4848]: E1204 15:02:26.402326 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:02:26 crc kubenswrapper[4848]: I1204 15:02:26.488819 4848 generic.go:334] "Generic (PLEG): container finished" podID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerID="9c6726243aa9ce482f6bef9410b748d2b9445fc325591a6cf970e2f1dbc83b9b" exitCode=0 Dec 04 15:02:26 crc kubenswrapper[4848]: I1204 15:02:26.489290 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2ntw" event={"ID":"455f8f2c-0179-4f4c-b5f7-584f0b091527","Type":"ContainerDied","Data":"9c6726243aa9ce482f6bef9410b748d2b9445fc325591a6cf970e2f1dbc83b9b"} Dec 04 15:02:26 crc kubenswrapper[4848]: I1204 15:02:26.874654 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.000503 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-utilities\") pod \"455f8f2c-0179-4f4c-b5f7-584f0b091527\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.000656 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x76x2\" (UniqueName: \"kubernetes.io/projected/455f8f2c-0179-4f4c-b5f7-584f0b091527-kube-api-access-x76x2\") pod \"455f8f2c-0179-4f4c-b5f7-584f0b091527\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.000726 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-catalog-content\") pod \"455f8f2c-0179-4f4c-b5f7-584f0b091527\" (UID: \"455f8f2c-0179-4f4c-b5f7-584f0b091527\") " Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.001817 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-utilities" (OuterVolumeSpecName: "utilities") pod "455f8f2c-0179-4f4c-b5f7-584f0b091527" (UID: "455f8f2c-0179-4f4c-b5f7-584f0b091527"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.010437 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/455f8f2c-0179-4f4c-b5f7-584f0b091527-kube-api-access-x76x2" (OuterVolumeSpecName: "kube-api-access-x76x2") pod "455f8f2c-0179-4f4c-b5f7-584f0b091527" (UID: "455f8f2c-0179-4f4c-b5f7-584f0b091527"). InnerVolumeSpecName "kube-api-access-x76x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.103586 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.103887 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x76x2\" (UniqueName: \"kubernetes.io/projected/455f8f2c-0179-4f4c-b5f7-584f0b091527-kube-api-access-x76x2\") on node \"crc\" DevicePath \"\"" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.140455 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "455f8f2c-0179-4f4c-b5f7-584f0b091527" (UID: "455f8f2c-0179-4f4c-b5f7-584f0b091527"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.206166 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/455f8f2c-0179-4f4c-b5f7-584f0b091527-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.501504 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q2ntw" event={"ID":"455f8f2c-0179-4f4c-b5f7-584f0b091527","Type":"ContainerDied","Data":"94bcb927ab6f81176eac499f3d8a9af22e9043ff9f7b7f707f418bd1327084d1"} Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.501774 4848 scope.go:117] "RemoveContainer" containerID="9c6726243aa9ce482f6bef9410b748d2b9445fc325591a6cf970e2f1dbc83b9b" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.501573 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q2ntw" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.540601 4848 scope.go:117] "RemoveContainer" containerID="ed188a535dbf6f2864d97c002ddbb897b41c17a6810bda141862a29724d1936b" Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.542293 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q2ntw"] Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.561322 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q2ntw"] Dec 04 15:02:27 crc kubenswrapper[4848]: I1204 15:02:27.568104 4848 scope.go:117] "RemoveContainer" containerID="1a5846ee6fc7c1be56b34bbabec3d727df7b9bee399cebe406712ea65ee7121c" Dec 04 15:02:28 crc kubenswrapper[4848]: I1204 15:02:28.412666 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" path="/var/lib/kubelet/pods/455f8f2c-0179-4f4c-b5f7-584f0b091527/volumes" Dec 04 15:02:39 crc kubenswrapper[4848]: I1204 15:02:39.394212 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:02:39 crc kubenswrapper[4848]: E1204 15:02:39.395520 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:02:54 crc kubenswrapper[4848]: I1204 15:02:54.394619 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:02:54 crc kubenswrapper[4848]: E1204 15:02:54.398473 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:03:09 crc kubenswrapper[4848]: I1204 15:03:09.393709 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:03:09 crc kubenswrapper[4848]: E1204 15:03:09.394752 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:03:21 crc kubenswrapper[4848]: I1204 15:03:21.394724 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:03:21 crc kubenswrapper[4848]: E1204 15:03:21.396237 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:03:33 crc kubenswrapper[4848]: I1204 15:03:33.394134 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:03:33 crc kubenswrapper[4848]: E1204 15:03:33.395050 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.249276 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mjjhm"] Dec 04 15:03:39 crc kubenswrapper[4848]: E1204 15:03:39.250341 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="registry-server" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.250359 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="registry-server" Dec 04 15:03:39 crc kubenswrapper[4848]: E1204 15:03:39.250372 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="extract-content" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.250380 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="extract-content" Dec 04 15:03:39 crc kubenswrapper[4848]: E1204 15:03:39.250408 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="extract-utilities" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.250418 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="extract-utilities" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.250736 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="455f8f2c-0179-4f4c-b5f7-584f0b091527" containerName="registry-server" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.253213 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.267898 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mjjhm"] Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.410861 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-utilities\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.411002 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pplq\" (UniqueName: \"kubernetes.io/projected/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-kube-api-access-8pplq\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.411350 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-catalog-content\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.513278 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-utilities\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.513372 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pplq\" (UniqueName: \"kubernetes.io/projected/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-kube-api-access-8pplq\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.513472 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-catalog-content\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.513938 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-catalog-content\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.514489 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-utilities\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.533608 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pplq\" (UniqueName: \"kubernetes.io/projected/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-kube-api-access-8pplq\") pod \"community-operators-mjjhm\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:39 crc kubenswrapper[4848]: I1204 15:03:39.587193 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:40 crc kubenswrapper[4848]: I1204 15:03:40.148531 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mjjhm"] Dec 04 15:03:40 crc kubenswrapper[4848]: I1204 15:03:40.412709 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjhm" event={"ID":"bb49b3a8-983b-4c9d-81ce-e725933f3f7a","Type":"ContainerStarted","Data":"c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1"} Dec 04 15:03:40 crc kubenswrapper[4848]: I1204 15:03:40.413090 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjhm" event={"ID":"bb49b3a8-983b-4c9d-81ce-e725933f3f7a","Type":"ContainerStarted","Data":"e06a992a8c3957903bf7c228940d8d68cdf0eedce9dfd627fc84ce8006a0824f"} Dec 04 15:03:41 crc kubenswrapper[4848]: I1204 15:03:41.426401 4848 generic.go:334] "Generic (PLEG): container finished" podID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerID="c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1" exitCode=0 Dec 04 15:03:41 crc kubenswrapper[4848]: I1204 15:03:41.426443 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjhm" event={"ID":"bb49b3a8-983b-4c9d-81ce-e725933f3f7a","Type":"ContainerDied","Data":"c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1"} Dec 04 15:03:45 crc kubenswrapper[4848]: I1204 15:03:45.394453 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:03:45 crc kubenswrapper[4848]: E1204 15:03:45.395225 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:03:53 crc kubenswrapper[4848]: I1204 15:03:53.574768 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjhm" event={"ID":"bb49b3a8-983b-4c9d-81ce-e725933f3f7a","Type":"ContainerStarted","Data":"cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554"} Dec 04 15:03:54 crc kubenswrapper[4848]: I1204 15:03:54.586609 4848 generic.go:334] "Generic (PLEG): container finished" podID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerID="cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554" exitCode=0 Dec 04 15:03:54 crc kubenswrapper[4848]: I1204 15:03:54.586681 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjhm" event={"ID":"bb49b3a8-983b-4c9d-81ce-e725933f3f7a","Type":"ContainerDied","Data":"cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554"} Dec 04 15:03:55 crc kubenswrapper[4848]: I1204 15:03:55.602520 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjhm" event={"ID":"bb49b3a8-983b-4c9d-81ce-e725933f3f7a","Type":"ContainerStarted","Data":"c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a"} Dec 04 15:03:55 crc kubenswrapper[4848]: I1204 15:03:55.627718 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mjjhm" podStartSLOduration=2.825232186 podStartE2EDuration="16.627697416s" podCreationTimestamp="2025-12-04 15:03:39 +0000 UTC" firstStartedPulling="2025-12-04 15:03:41.429391402 +0000 UTC m=+4525.371887930" lastFinishedPulling="2025-12-04 15:03:55.231856632 +0000 UTC m=+4539.174353160" observedRunningTime="2025-12-04 15:03:55.624076377 +0000 UTC m=+4539.566572915" watchObservedRunningTime="2025-12-04 15:03:55.627697416 +0000 UTC m=+4539.570193944" Dec 04 15:03:56 crc kubenswrapper[4848]: I1204 15:03:56.402172 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:03:56 crc kubenswrapper[4848]: E1204 15:03:56.402795 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:03:59 crc kubenswrapper[4848]: I1204 15:03:59.587832 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:59 crc kubenswrapper[4848]: I1204 15:03:59.588485 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:03:59 crc kubenswrapper[4848]: I1204 15:03:59.648872 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:04:07 crc kubenswrapper[4848]: I1204 15:04:07.394567 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:04:07 crc kubenswrapper[4848]: E1204 15:04:07.395552 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:04:09 crc kubenswrapper[4848]: I1204 15:04:09.660670 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:04:09 crc kubenswrapper[4848]: I1204 15:04:09.734058 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mjjhm"] Dec 04 15:04:09 crc kubenswrapper[4848]: I1204 15:04:09.773115 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mjjhm" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerName="registry-server" containerID="cri-o://c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a" gracePeriod=2 Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.373449 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.499695 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-catalog-content\") pod \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.500229 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-utilities\") pod \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.500297 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pplq\" (UniqueName: \"kubernetes.io/projected/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-kube-api-access-8pplq\") pod \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\" (UID: \"bb49b3a8-983b-4c9d-81ce-e725933f3f7a\") " Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.500839 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-utilities" (OuterVolumeSpecName: "utilities") pod "bb49b3a8-983b-4c9d-81ce-e725933f3f7a" (UID: "bb49b3a8-983b-4c9d-81ce-e725933f3f7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.502366 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.508173 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-kube-api-access-8pplq" (OuterVolumeSpecName: "kube-api-access-8pplq") pod "bb49b3a8-983b-4c9d-81ce-e725933f3f7a" (UID: "bb49b3a8-983b-4c9d-81ce-e725933f3f7a"). InnerVolumeSpecName "kube-api-access-8pplq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.554757 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb49b3a8-983b-4c9d-81ce-e725933f3f7a" (UID: "bb49b3a8-983b-4c9d-81ce-e725933f3f7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.605033 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.605063 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pplq\" (UniqueName: \"kubernetes.io/projected/bb49b3a8-983b-4c9d-81ce-e725933f3f7a-kube-api-access-8pplq\") on node \"crc\" DevicePath \"\"" Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.785355 4848 generic.go:334] "Generic (PLEG): container finished" podID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerID="c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a" exitCode=0 Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.785404 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjhm" event={"ID":"bb49b3a8-983b-4c9d-81ce-e725933f3f7a","Type":"ContainerDied","Data":"c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a"} Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.785436 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mjjhm" event={"ID":"bb49b3a8-983b-4c9d-81ce-e725933f3f7a","Type":"ContainerDied","Data":"e06a992a8c3957903bf7c228940d8d68cdf0eedce9dfd627fc84ce8006a0824f"} Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.785449 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mjjhm" Dec 04 15:04:10 crc kubenswrapper[4848]: I1204 15:04:10.785457 4848 scope.go:117] "RemoveContainer" containerID="c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a" Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.250857 4848 scope.go:117] "RemoveContainer" containerID="cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554" Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.255063 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mjjhm"] Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.271780 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mjjhm"] Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.356125 4848 scope.go:117] "RemoveContainer" containerID="c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1" Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.395388 4848 scope.go:117] "RemoveContainer" containerID="c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a" Dec 04 15:04:11 crc kubenswrapper[4848]: E1204 15:04:11.396003 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a\": container with ID starting with c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a not found: ID does not exist" containerID="c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a" Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.396050 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a"} err="failed to get container status \"c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a\": rpc error: code = NotFound desc = could not find container \"c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a\": container with ID starting with c1bf36f011cb78f950b4030d24283544fa5c85a11312273aa960df8cbc83484a not found: ID does not exist" Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.396075 4848 scope.go:117] "RemoveContainer" containerID="cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554" Dec 04 15:04:11 crc kubenswrapper[4848]: E1204 15:04:11.396500 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554\": container with ID starting with cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554 not found: ID does not exist" containerID="cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554" Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.396537 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554"} err="failed to get container status \"cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554\": rpc error: code = NotFound desc = could not find container \"cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554\": container with ID starting with cae5c859c53d7e694102fd689cd516e13dd2c07fd0f9b3c95936cbc21823b554 not found: ID does not exist" Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.396564 4848 scope.go:117] "RemoveContainer" containerID="c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1" Dec 04 15:04:11 crc kubenswrapper[4848]: E1204 15:04:11.396885 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1\": container with ID starting with c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1 not found: ID does not exist" containerID="c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1" Dec 04 15:04:11 crc kubenswrapper[4848]: I1204 15:04:11.396919 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1"} err="failed to get container status \"c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1\": rpc error: code = NotFound desc = could not find container \"c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1\": container with ID starting with c23e8bf71cd648ce2bc521de8583c045fcdf4eb2dae267a0b4c213fbb58448a1 not found: ID does not exist" Dec 04 15:04:12 crc kubenswrapper[4848]: I1204 15:04:12.407873 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" path="/var/lib/kubelet/pods/bb49b3a8-983b-4c9d-81ce-e725933f3f7a/volumes" Dec 04 15:04:19 crc kubenswrapper[4848]: I1204 15:04:19.394254 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:04:19 crc kubenswrapper[4848]: E1204 15:04:19.395302 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:04:32 crc kubenswrapper[4848]: I1204 15:04:32.393862 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:04:32 crc kubenswrapper[4848]: E1204 15:04:32.394753 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:04:44 crc kubenswrapper[4848]: I1204 15:04:44.394118 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:04:44 crc kubenswrapper[4848]: E1204 15:04:44.394672 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:04:50 crc kubenswrapper[4848]: E1204 15:04:50.405787 4848 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.5:59334->38.102.83.5:35999: write tcp 38.102.83.5:59334->38.102.83.5:35999: write: broken pipe Dec 04 15:04:55 crc kubenswrapper[4848]: I1204 15:04:55.394201 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:04:55 crc kubenswrapper[4848]: E1204 15:04:55.395158 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:05:09 crc kubenswrapper[4848]: I1204 15:05:09.393349 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:05:09 crc kubenswrapper[4848]: E1204 15:05:09.394227 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:05:22 crc kubenswrapper[4848]: I1204 15:05:22.394385 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:05:22 crc kubenswrapper[4848]: E1204 15:05:22.395256 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:05:34 crc kubenswrapper[4848]: I1204 15:05:34.394075 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:05:34 crc kubenswrapper[4848]: E1204 15:05:34.395041 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:05:49 crc kubenswrapper[4848]: I1204 15:05:49.393876 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:05:49 crc kubenswrapper[4848]: E1204 15:05:49.394883 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:06:01 crc kubenswrapper[4848]: I1204 15:06:01.393771 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:06:01 crc kubenswrapper[4848]: E1204 15:06:01.394774 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:06:13 crc kubenswrapper[4848]: I1204 15:06:13.394095 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:06:13 crc kubenswrapper[4848]: E1204 15:06:13.394865 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:06:25 crc kubenswrapper[4848]: I1204 15:06:25.394168 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:06:25 crc kubenswrapper[4848]: E1204 15:06:25.395066 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:06:39 crc kubenswrapper[4848]: I1204 15:06:39.393310 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:06:39 crc kubenswrapper[4848]: E1204 15:06:39.394435 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:06:52 crc kubenswrapper[4848]: I1204 15:06:52.394849 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:06:53 crc kubenswrapper[4848]: I1204 15:06:53.559875 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"f8c2293d3571233fe85bf452139bb0b6575ffc1fa006ccbe36464f352d131f7a"} Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.466469 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 15:08:50 crc kubenswrapper[4848]: E1204 15:08:50.467712 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerName="extract-content" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.467737 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerName="extract-content" Dec 04 15:08:50 crc kubenswrapper[4848]: E1204 15:08:50.467758 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerName="extract-utilities" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.467773 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerName="extract-utilities" Dec 04 15:08:50 crc kubenswrapper[4848]: E1204 15:08:50.467843 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerName="registry-server" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.467857 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerName="registry-server" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.468317 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb49b3a8-983b-4c9d-81ce-e725933f3f7a" containerName="registry-server" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.469634 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.473500 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.476158 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kfvbp" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.476328 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.476466 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.482324 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.488109 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-config-data\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.488160 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.488429 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.590496 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.591238 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.591377 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.591592 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2thsw\" (UniqueName: \"kubernetes.io/projected/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-kube-api-access-2thsw\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.591765 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.591867 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.592018 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.592170 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-config-data\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.592292 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.594597 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.595084 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-config-data\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.598217 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.694284 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2thsw\" (UniqueName: \"kubernetes.io/projected/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-kube-api-access-2thsw\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.694690 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.694774 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.694937 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.695264 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.695440 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.695495 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.695520 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.697916 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.699389 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.699901 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.713635 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2thsw\" (UniqueName: \"kubernetes.io/projected/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-kube-api-access-2thsw\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.735669 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " pod="openstack/tempest-tests-tempest" Dec 04 15:08:50 crc kubenswrapper[4848]: I1204 15:08:50.793940 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 15:08:51 crc kubenswrapper[4848]: I1204 15:08:51.388710 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 15:08:51 crc kubenswrapper[4848]: I1204 15:08:51.394387 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:08:52 crc kubenswrapper[4848]: I1204 15:08:52.305275 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a","Type":"ContainerStarted","Data":"046bd5f1ac0c0f08de9648b6733b739264c1de352b8a16ec2fe6994ddda07bc7"} Dec 04 15:09:14 crc kubenswrapper[4848]: I1204 15:09:14.314768 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:09:14 crc kubenswrapper[4848]: I1204 15:09:14.315364 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:09:29 crc kubenswrapper[4848]: E1204 15:09:29.850712 4848 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 04 15:09:29 crc kubenswrapper[4848]: E1204 15:09:29.852594 4848 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2thsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 15:09:29 crc kubenswrapper[4848]: E1204 15:09:29.853702 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" Dec 04 15:09:30 crc kubenswrapper[4848]: E1204 15:09:30.733103 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" Dec 04 15:09:42 crc kubenswrapper[4848]: I1204 15:09:42.974487 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 04 15:09:44 crc kubenswrapper[4848]: I1204 15:09:44.314803 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:09:44 crc kubenswrapper[4848]: I1204 15:09:44.315441 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:09:44 crc kubenswrapper[4848]: I1204 15:09:44.885933 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a","Type":"ContainerStarted","Data":"290ba3831b53925b3626369bf155fb33b9789c8116171885778ba9c197cae4c3"} Dec 04 15:09:44 crc kubenswrapper[4848]: I1204 15:09:44.911604 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.334213211 podStartE2EDuration="55.911585399s" podCreationTimestamp="2025-12-04 15:08:49 +0000 UTC" firstStartedPulling="2025-12-04 15:08:51.394109695 +0000 UTC m=+4835.336606223" lastFinishedPulling="2025-12-04 15:09:42.971481883 +0000 UTC m=+4886.913978411" observedRunningTime="2025-12-04 15:09:44.901546492 +0000 UTC m=+4888.844043020" watchObservedRunningTime="2025-12-04 15:09:44.911585399 +0000 UTC m=+4888.854081927" Dec 04 15:10:14 crc kubenswrapper[4848]: I1204 15:10:14.314691 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:10:14 crc kubenswrapper[4848]: I1204 15:10:14.315302 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:10:14 crc kubenswrapper[4848]: I1204 15:10:14.315361 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:10:14 crc kubenswrapper[4848]: I1204 15:10:14.316618 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8c2293d3571233fe85bf452139bb0b6575ffc1fa006ccbe36464f352d131f7a"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:10:14 crc kubenswrapper[4848]: I1204 15:10:14.316697 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://f8c2293d3571233fe85bf452139bb0b6575ffc1fa006ccbe36464f352d131f7a" gracePeriod=600 Dec 04 15:10:15 crc kubenswrapper[4848]: I1204 15:10:15.244580 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="f8c2293d3571233fe85bf452139bb0b6575ffc1fa006ccbe36464f352d131f7a" exitCode=0 Dec 04 15:10:15 crc kubenswrapper[4848]: I1204 15:10:15.244675 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"f8c2293d3571233fe85bf452139bb0b6575ffc1fa006ccbe36464f352d131f7a"} Dec 04 15:10:15 crc kubenswrapper[4848]: I1204 15:10:15.245146 4848 scope.go:117] "RemoveContainer" containerID="2327051091fba3ffd0e033380ddde2912754895aa9e7a39c8602d60430da4404" Dec 04 15:10:16 crc kubenswrapper[4848]: I1204 15:10:16.258795 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba"} Dec 04 15:12:44 crc kubenswrapper[4848]: I1204 15:12:44.314607 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:12:44 crc kubenswrapper[4848]: I1204 15:12:44.315159 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.005607 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9wnx8"] Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.015638 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.025906 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9wnx8"] Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.207562 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxh7x\" (UniqueName: \"kubernetes.io/projected/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-kube-api-access-lxh7x\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.207659 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-catalog-content\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.208748 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-utilities\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.310187 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-utilities\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.310315 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxh7x\" (UniqueName: \"kubernetes.io/projected/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-kube-api-access-lxh7x\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.310370 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-catalog-content\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.312732 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-utilities\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.313264 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-catalog-content\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.336982 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxh7x\" (UniqueName: \"kubernetes.io/projected/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-kube-api-access-lxh7x\") pod \"redhat-operators-9wnx8\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:10 crc kubenswrapper[4848]: I1204 15:13:10.346170 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:11 crc kubenswrapper[4848]: I1204 15:13:11.328868 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9wnx8"] Dec 04 15:13:11 crc kubenswrapper[4848]: I1204 15:13:11.904698 4848 generic.go:334] "Generic (PLEG): container finished" podID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerID="53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e" exitCode=0 Dec 04 15:13:11 crc kubenswrapper[4848]: I1204 15:13:11.904909 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnx8" event={"ID":"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d","Type":"ContainerDied","Data":"53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e"} Dec 04 15:13:11 crc kubenswrapper[4848]: I1204 15:13:11.905578 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnx8" event={"ID":"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d","Type":"ContainerStarted","Data":"0f17430350a72321145564417dacffcd664baeae54d133021ab9b1a8faaeaf55"} Dec 04 15:13:12 crc kubenswrapper[4848]: I1204 15:13:12.921009 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnx8" event={"ID":"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d","Type":"ContainerStarted","Data":"f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8"} Dec 04 15:13:14 crc kubenswrapper[4848]: I1204 15:13:14.313837 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:13:14 crc kubenswrapper[4848]: I1204 15:13:14.313897 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:13:16 crc kubenswrapper[4848]: I1204 15:13:16.993100 4848 generic.go:334] "Generic (PLEG): container finished" podID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerID="f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8" exitCode=0 Dec 04 15:13:16 crc kubenswrapper[4848]: I1204 15:13:16.993714 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnx8" event={"ID":"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d","Type":"ContainerDied","Data":"f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8"} Dec 04 15:13:18 crc kubenswrapper[4848]: I1204 15:13:18.007300 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnx8" event={"ID":"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d","Type":"ContainerStarted","Data":"9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969"} Dec 04 15:13:18 crc kubenswrapper[4848]: I1204 15:13:18.033325 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9wnx8" podStartSLOduration=3.531507313 podStartE2EDuration="9.033293253s" podCreationTimestamp="2025-12-04 15:13:09 +0000 UTC" firstStartedPulling="2025-12-04 15:13:11.906775045 +0000 UTC m=+5095.849271573" lastFinishedPulling="2025-12-04 15:13:17.408560985 +0000 UTC m=+5101.351057513" observedRunningTime="2025-12-04 15:13:18.025792248 +0000 UTC m=+5101.968288776" watchObservedRunningTime="2025-12-04 15:13:18.033293253 +0000 UTC m=+5101.975789771" Dec 04 15:13:20 crc kubenswrapper[4848]: I1204 15:13:20.346941 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:20 crc kubenswrapper[4848]: I1204 15:13:20.347493 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:21 crc kubenswrapper[4848]: I1204 15:13:21.396552 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9wnx8" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="registry-server" probeResult="failure" output=< Dec 04 15:13:21 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:13:21 crc kubenswrapper[4848]: > Dec 04 15:13:31 crc kubenswrapper[4848]: I1204 15:13:31.492683 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9wnx8" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="registry-server" probeResult="failure" output=< Dec 04 15:13:31 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:13:31 crc kubenswrapper[4848]: > Dec 04 15:13:40 crc kubenswrapper[4848]: I1204 15:13:40.411766 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:40 crc kubenswrapper[4848]: I1204 15:13:40.466713 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:41 crc kubenswrapper[4848]: I1204 15:13:41.137917 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9wnx8"] Dec 04 15:13:41 crc kubenswrapper[4848]: I1204 15:13:41.744372 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2" containerName="galera" probeResult="failure" output="command timed out" Dec 04 15:13:41 crc kubenswrapper[4848]: I1204 15:13:41.744391 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2" containerName="galera" probeResult="failure" output="command timed out" Dec 04 15:13:42 crc kubenswrapper[4848]: I1204 15:13:42.302526 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9wnx8" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="registry-server" containerID="cri-o://9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969" gracePeriod=2 Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.007979 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.114906 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-catalog-content\") pod \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.115466 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-utilities\") pod \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.115543 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxh7x\" (UniqueName: \"kubernetes.io/projected/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-kube-api-access-lxh7x\") pod \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\" (UID: \"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d\") " Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.118656 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-utilities" (OuterVolumeSpecName: "utilities") pod "fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" (UID: "fbd8a05f-16d5-4385-9ab2-66c55b3ab99d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.137293 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-kube-api-access-lxh7x" (OuterVolumeSpecName: "kube-api-access-lxh7x") pod "fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" (UID: "fbd8a05f-16d5-4385-9ab2-66c55b3ab99d"). InnerVolumeSpecName "kube-api-access-lxh7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.219118 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.219162 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxh7x\" (UniqueName: \"kubernetes.io/projected/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-kube-api-access-lxh7x\") on node \"crc\" DevicePath \"\"" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.228599 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" (UID: "fbd8a05f-16d5-4385-9ab2-66c55b3ab99d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.313503 4848 generic.go:334] "Generic (PLEG): container finished" podID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerID="9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969" exitCode=0 Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.313556 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnx8" event={"ID":"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d","Type":"ContainerDied","Data":"9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969"} Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.313588 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9wnx8" event={"ID":"fbd8a05f-16d5-4385-9ab2-66c55b3ab99d","Type":"ContainerDied","Data":"0f17430350a72321145564417dacffcd664baeae54d133021ab9b1a8faaeaf55"} Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.313611 4848 scope.go:117] "RemoveContainer" containerID="9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.313559 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9wnx8" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.322472 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.353737 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9wnx8"] Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.370107 4848 scope.go:117] "RemoveContainer" containerID="f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.373129 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9wnx8"] Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.391913 4848 scope.go:117] "RemoveContainer" containerID="53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.444302 4848 scope.go:117] "RemoveContainer" containerID="9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969" Dec 04 15:13:43 crc kubenswrapper[4848]: E1204 15:13:43.446117 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969\": container with ID starting with 9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969 not found: ID does not exist" containerID="9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.446345 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969"} err="failed to get container status \"9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969\": rpc error: code = NotFound desc = could not find container \"9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969\": container with ID starting with 9f6db65bd0619fda24ae2cd9fa93b164db67e0a37417e3cd3219ac3f6009f969 not found: ID does not exist" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.446380 4848 scope.go:117] "RemoveContainer" containerID="f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8" Dec 04 15:13:43 crc kubenswrapper[4848]: E1204 15:13:43.446816 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8\": container with ID starting with f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8 not found: ID does not exist" containerID="f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.446853 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8"} err="failed to get container status \"f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8\": rpc error: code = NotFound desc = could not find container \"f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8\": container with ID starting with f93a78529651cc9a72c7308647ef69f805f45cd44a1cd9dce1ae81c41cd43da8 not found: ID does not exist" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.446878 4848 scope.go:117] "RemoveContainer" containerID="53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e" Dec 04 15:13:43 crc kubenswrapper[4848]: E1204 15:13:43.447217 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e\": container with ID starting with 53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e not found: ID does not exist" containerID="53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e" Dec 04 15:13:43 crc kubenswrapper[4848]: I1204 15:13:43.447245 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e"} err="failed to get container status \"53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e\": rpc error: code = NotFound desc = could not find container \"53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e\": container with ID starting with 53f396160a396909b4979f429f9bf4f849da646837473e5579afb0c79641db2e not found: ID does not exist" Dec 04 15:13:44 crc kubenswrapper[4848]: I1204 15:13:44.314418 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:13:44 crc kubenswrapper[4848]: I1204 15:13:44.314498 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:13:44 crc kubenswrapper[4848]: I1204 15:13:44.314605 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:13:44 crc kubenswrapper[4848]: I1204 15:13:44.315943 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:13:44 crc kubenswrapper[4848]: I1204 15:13:44.316055 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" gracePeriod=600 Dec 04 15:13:44 crc kubenswrapper[4848]: I1204 15:13:44.411408 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" path="/var/lib/kubelet/pods/fbd8a05f-16d5-4385-9ab2-66c55b3ab99d/volumes" Dec 04 15:13:44 crc kubenswrapper[4848]: E1204 15:13:44.438217 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:13:45 crc kubenswrapper[4848]: I1204 15:13:45.408249 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" exitCode=0 Dec 04 15:13:45 crc kubenswrapper[4848]: I1204 15:13:45.408299 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba"} Dec 04 15:13:45 crc kubenswrapper[4848]: I1204 15:13:45.408335 4848 scope.go:117] "RemoveContainer" containerID="f8c2293d3571233fe85bf452139bb0b6575ffc1fa006ccbe36464f352d131f7a" Dec 04 15:13:45 crc kubenswrapper[4848]: I1204 15:13:45.409410 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:13:45 crc kubenswrapper[4848]: E1204 15:13:45.409816 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:14:00 crc kubenswrapper[4848]: I1204 15:14:00.393608 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:14:00 crc kubenswrapper[4848]: E1204 15:14:00.394573 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:14:13 crc kubenswrapper[4848]: I1204 15:14:13.394338 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:14:13 crc kubenswrapper[4848]: E1204 15:14:13.395147 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:14:24 crc kubenswrapper[4848]: I1204 15:14:24.394746 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:14:24 crc kubenswrapper[4848]: E1204 15:14:24.395426 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.172170 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wn2nn"] Dec 04 15:14:29 crc kubenswrapper[4848]: E1204 15:14:29.173406 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="registry-server" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.173609 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="registry-server" Dec 04 15:14:29 crc kubenswrapper[4848]: E1204 15:14:29.173652 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="extract-content" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.173659 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="extract-content" Dec 04 15:14:29 crc kubenswrapper[4848]: E1204 15:14:29.173686 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="extract-utilities" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.173693 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="extract-utilities" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.174043 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbd8a05f-16d5-4385-9ab2-66c55b3ab99d" containerName="registry-server" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.175677 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.211638 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wn2nn"] Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.227041 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-utilities\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.227156 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-catalog-content\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.227203 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfx7m\" (UniqueName: \"kubernetes.io/projected/935f2c00-30f1-45ed-9934-4f8b26d1f045-kube-api-access-cfx7m\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.329568 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-utilities\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.329710 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-catalog-content\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.329741 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfx7m\" (UniqueName: \"kubernetes.io/projected/935f2c00-30f1-45ed-9934-4f8b26d1f045-kube-api-access-cfx7m\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.330077 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-utilities\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.330303 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-catalog-content\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.362258 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfx7m\" (UniqueName: \"kubernetes.io/projected/935f2c00-30f1-45ed-9934-4f8b26d1f045-kube-api-access-cfx7m\") pod \"community-operators-wn2nn\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:29 crc kubenswrapper[4848]: I1204 15:14:29.512410 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:30 crc kubenswrapper[4848]: I1204 15:14:30.049845 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wn2nn"] Dec 04 15:14:30 crc kubenswrapper[4848]: I1204 15:14:30.892637 4848 generic.go:334] "Generic (PLEG): container finished" podID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerID="639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841" exitCode=0 Dec 04 15:14:30 crc kubenswrapper[4848]: I1204 15:14:30.892822 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wn2nn" event={"ID":"935f2c00-30f1-45ed-9934-4f8b26d1f045","Type":"ContainerDied","Data":"639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841"} Dec 04 15:14:30 crc kubenswrapper[4848]: I1204 15:14:30.894023 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wn2nn" event={"ID":"935f2c00-30f1-45ed-9934-4f8b26d1f045","Type":"ContainerStarted","Data":"197ae229b51021a639a253036fd21f5308daadd2be7427c4c2ed30d6a72fd5a2"} Dec 04 15:14:30 crc kubenswrapper[4848]: I1204 15:14:30.895665 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:14:32 crc kubenswrapper[4848]: I1204 15:14:32.915819 4848 generic.go:334] "Generic (PLEG): container finished" podID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerID="048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719" exitCode=0 Dec 04 15:14:32 crc kubenswrapper[4848]: I1204 15:14:32.915881 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wn2nn" event={"ID":"935f2c00-30f1-45ed-9934-4f8b26d1f045","Type":"ContainerDied","Data":"048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719"} Dec 04 15:14:34 crc kubenswrapper[4848]: I1204 15:14:34.995921 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wn2nn" event={"ID":"935f2c00-30f1-45ed-9934-4f8b26d1f045","Type":"ContainerStarted","Data":"e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d"} Dec 04 15:14:35 crc kubenswrapper[4848]: I1204 15:14:35.026091 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wn2nn" podStartSLOduration=3.270911836 podStartE2EDuration="6.02607244s" podCreationTimestamp="2025-12-04 15:14:29 +0000 UTC" firstStartedPulling="2025-12-04 15:14:30.894982549 +0000 UTC m=+5174.837479087" lastFinishedPulling="2025-12-04 15:14:33.650143163 +0000 UTC m=+5177.592639691" observedRunningTime="2025-12-04 15:14:35.020462672 +0000 UTC m=+5178.962959200" watchObservedRunningTime="2025-12-04 15:14:35.02607244 +0000 UTC m=+5178.968568968" Dec 04 15:14:38 crc kubenswrapper[4848]: I1204 15:14:38.394156 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:14:38 crc kubenswrapper[4848]: E1204 15:14:38.394861 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:14:39 crc kubenswrapper[4848]: I1204 15:14:39.513180 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:39 crc kubenswrapper[4848]: I1204 15:14:39.964853 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:39 crc kubenswrapper[4848]: I1204 15:14:39.970241 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:40 crc kubenswrapper[4848]: I1204 15:14:40.406705 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:40 crc kubenswrapper[4848]: I1204 15:14:40.455197 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wn2nn"] Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.069431 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wn2nn" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerName="registry-server" containerID="cri-o://e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d" gracePeriod=2 Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.665108 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.843754 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfx7m\" (UniqueName: \"kubernetes.io/projected/935f2c00-30f1-45ed-9934-4f8b26d1f045-kube-api-access-cfx7m\") pod \"935f2c00-30f1-45ed-9934-4f8b26d1f045\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.843865 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-catalog-content\") pod \"935f2c00-30f1-45ed-9934-4f8b26d1f045\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.843908 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-utilities\") pod \"935f2c00-30f1-45ed-9934-4f8b26d1f045\" (UID: \"935f2c00-30f1-45ed-9934-4f8b26d1f045\") " Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.844726 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-utilities" (OuterVolumeSpecName: "utilities") pod "935f2c00-30f1-45ed-9934-4f8b26d1f045" (UID: "935f2c00-30f1-45ed-9934-4f8b26d1f045"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.851439 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/935f2c00-30f1-45ed-9934-4f8b26d1f045-kube-api-access-cfx7m" (OuterVolumeSpecName: "kube-api-access-cfx7m") pod "935f2c00-30f1-45ed-9934-4f8b26d1f045" (UID: "935f2c00-30f1-45ed-9934-4f8b26d1f045"). InnerVolumeSpecName "kube-api-access-cfx7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.895667 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "935f2c00-30f1-45ed-9934-4f8b26d1f045" (UID: "935f2c00-30f1-45ed-9934-4f8b26d1f045"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.947145 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.947178 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/935f2c00-30f1-45ed-9934-4f8b26d1f045-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:14:42 crc kubenswrapper[4848]: I1204 15:14:42.947188 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfx7m\" (UniqueName: \"kubernetes.io/projected/935f2c00-30f1-45ed-9934-4f8b26d1f045-kube-api-access-cfx7m\") on node \"crc\" DevicePath \"\"" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.082033 4848 generic.go:334] "Generic (PLEG): container finished" podID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerID="e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d" exitCode=0 Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.082084 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wn2nn" event={"ID":"935f2c00-30f1-45ed-9934-4f8b26d1f045","Type":"ContainerDied","Data":"e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d"} Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.082117 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wn2nn" event={"ID":"935f2c00-30f1-45ed-9934-4f8b26d1f045","Type":"ContainerDied","Data":"197ae229b51021a639a253036fd21f5308daadd2be7427c4c2ed30d6a72fd5a2"} Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.082138 4848 scope.go:117] "RemoveContainer" containerID="e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.082192 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wn2nn" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.131031 4848 scope.go:117] "RemoveContainer" containerID="048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.166844 4848 scope.go:117] "RemoveContainer" containerID="639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.169458 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wn2nn"] Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.181490 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wn2nn"] Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.212918 4848 scope.go:117] "RemoveContainer" containerID="e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d" Dec 04 15:14:43 crc kubenswrapper[4848]: E1204 15:14:43.213540 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d\": container with ID starting with e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d not found: ID does not exist" containerID="e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.213584 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d"} err="failed to get container status \"e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d\": rpc error: code = NotFound desc = could not find container \"e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d\": container with ID starting with e015bace55401c8ad026f036c9a33836566cd095890240af8ac1386b0a5ca68d not found: ID does not exist" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.213612 4848 scope.go:117] "RemoveContainer" containerID="048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719" Dec 04 15:14:43 crc kubenswrapper[4848]: E1204 15:14:43.213923 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719\": container with ID starting with 048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719 not found: ID does not exist" containerID="048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.213981 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719"} err="failed to get container status \"048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719\": rpc error: code = NotFound desc = could not find container \"048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719\": container with ID starting with 048ace6c51cd5d1477abec7331f87451a9e453b7cfe216d2fb18789f2b424719 not found: ID does not exist" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.214002 4848 scope.go:117] "RemoveContainer" containerID="639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841" Dec 04 15:14:43 crc kubenswrapper[4848]: E1204 15:14:43.214275 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841\": container with ID starting with 639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841 not found: ID does not exist" containerID="639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841" Dec 04 15:14:43 crc kubenswrapper[4848]: I1204 15:14:43.214305 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841"} err="failed to get container status \"639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841\": rpc error: code = NotFound desc = could not find container \"639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841\": container with ID starting with 639d0ccc45c22d08a47702267bff5e5adf77c874e613a8ce6a0e064e0789c841 not found: ID does not exist" Dec 04 15:14:44 crc kubenswrapper[4848]: I1204 15:14:44.421924 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" path="/var/lib/kubelet/pods/935f2c00-30f1-45ed-9934-4f8b26d1f045/volumes" Dec 04 15:14:52 crc kubenswrapper[4848]: I1204 15:14:52.394299 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:14:52 crc kubenswrapper[4848]: E1204 15:14:52.395069 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.183390 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt"] Dec 04 15:15:00 crc kubenswrapper[4848]: E1204 15:15:00.184679 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerName="extract-content" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.184703 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerName="extract-content" Dec 04 15:15:00 crc kubenswrapper[4848]: E1204 15:15:00.184724 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.184734 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4848]: E1204 15:15:00.184771 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerName="extract-utilities" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.184784 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerName="extract-utilities" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.185217 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="935f2c00-30f1-45ed-9934-4f8b26d1f045" containerName="registry-server" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.186454 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.199854 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.200241 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.209048 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt"] Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.365085 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fb1088d-906b-41ed-acd5-dc80377be4d4-config-volume\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.365146 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c49rc\" (UniqueName: \"kubernetes.io/projected/6fb1088d-906b-41ed-acd5-dc80377be4d4-kube-api-access-c49rc\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.365183 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fb1088d-906b-41ed-acd5-dc80377be4d4-secret-volume\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.467624 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fb1088d-906b-41ed-acd5-dc80377be4d4-config-volume\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.467687 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c49rc\" (UniqueName: \"kubernetes.io/projected/6fb1088d-906b-41ed-acd5-dc80377be4d4-kube-api-access-c49rc\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.467774 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fb1088d-906b-41ed-acd5-dc80377be4d4-secret-volume\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.470257 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fb1088d-906b-41ed-acd5-dc80377be4d4-config-volume\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.478515 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fb1088d-906b-41ed-acd5-dc80377be4d4-secret-volume\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.503168 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c49rc\" (UniqueName: \"kubernetes.io/projected/6fb1088d-906b-41ed-acd5-dc80377be4d4-kube-api-access-c49rc\") pod \"collect-profiles-29414355-66rtt\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:00 crc kubenswrapper[4848]: I1204 15:15:00.512425 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:01 crc kubenswrapper[4848]: I1204 15:15:01.001601 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt"] Dec 04 15:15:01 crc kubenswrapper[4848]: I1204 15:15:01.278404 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" event={"ID":"6fb1088d-906b-41ed-acd5-dc80377be4d4","Type":"ContainerStarted","Data":"b5dc7358389230150294b3f061c0123c85f4f398bb8cc7902b9b3fe98775004d"} Dec 04 15:15:01 crc kubenswrapper[4848]: I1204 15:15:01.278791 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" event={"ID":"6fb1088d-906b-41ed-acd5-dc80377be4d4","Type":"ContainerStarted","Data":"53eaa9d7fefbad981855ccdf51414990a720d32681db65416c6bb1d3c47f235f"} Dec 04 15:15:01 crc kubenswrapper[4848]: I1204 15:15:01.301781 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" podStartSLOduration=1.301759546 podStartE2EDuration="1.301759546s" podCreationTimestamp="2025-12-04 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 15:15:01.2990215 +0000 UTC m=+5205.241518048" watchObservedRunningTime="2025-12-04 15:15:01.301759546 +0000 UTC m=+5205.244256074" Dec 04 15:15:02 crc kubenswrapper[4848]: I1204 15:15:02.290209 4848 generic.go:334] "Generic (PLEG): container finished" podID="6fb1088d-906b-41ed-acd5-dc80377be4d4" containerID="b5dc7358389230150294b3f061c0123c85f4f398bb8cc7902b9b3fe98775004d" exitCode=0 Dec 04 15:15:02 crc kubenswrapper[4848]: I1204 15:15:02.290432 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" event={"ID":"6fb1088d-906b-41ed-acd5-dc80377be4d4","Type":"ContainerDied","Data":"b5dc7358389230150294b3f061c0123c85f4f398bb8cc7902b9b3fe98775004d"} Dec 04 15:15:03 crc kubenswrapper[4848]: I1204 15:15:03.823598 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:03 crc kubenswrapper[4848]: I1204 15:15:03.973442 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fb1088d-906b-41ed-acd5-dc80377be4d4-config-volume\") pod \"6fb1088d-906b-41ed-acd5-dc80377be4d4\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " Dec 04 15:15:03 crc kubenswrapper[4848]: I1204 15:15:03.973874 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c49rc\" (UniqueName: \"kubernetes.io/projected/6fb1088d-906b-41ed-acd5-dc80377be4d4-kube-api-access-c49rc\") pod \"6fb1088d-906b-41ed-acd5-dc80377be4d4\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " Dec 04 15:15:03 crc kubenswrapper[4848]: I1204 15:15:03.973959 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fb1088d-906b-41ed-acd5-dc80377be4d4-secret-volume\") pod \"6fb1088d-906b-41ed-acd5-dc80377be4d4\" (UID: \"6fb1088d-906b-41ed-acd5-dc80377be4d4\") " Dec 04 15:15:03 crc kubenswrapper[4848]: I1204 15:15:03.974300 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fb1088d-906b-41ed-acd5-dc80377be4d4-config-volume" (OuterVolumeSpecName: "config-volume") pod "6fb1088d-906b-41ed-acd5-dc80377be4d4" (UID: "6fb1088d-906b-41ed-acd5-dc80377be4d4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 15:15:03 crc kubenswrapper[4848]: I1204 15:15:03.974880 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fb1088d-906b-41ed-acd5-dc80377be4d4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:03 crc kubenswrapper[4848]: I1204 15:15:03.980350 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fb1088d-906b-41ed-acd5-dc80377be4d4-kube-api-access-c49rc" (OuterVolumeSpecName: "kube-api-access-c49rc") pod "6fb1088d-906b-41ed-acd5-dc80377be4d4" (UID: "6fb1088d-906b-41ed-acd5-dc80377be4d4"). InnerVolumeSpecName "kube-api-access-c49rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:15:03 crc kubenswrapper[4848]: I1204 15:15:03.981496 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb1088d-906b-41ed-acd5-dc80377be4d4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6fb1088d-906b-41ed-acd5-dc80377be4d4" (UID: "6fb1088d-906b-41ed-acd5-dc80377be4d4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:15:04 crc kubenswrapper[4848]: I1204 15:15:04.076642 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c49rc\" (UniqueName: \"kubernetes.io/projected/6fb1088d-906b-41ed-acd5-dc80377be4d4-kube-api-access-c49rc\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:04 crc kubenswrapper[4848]: I1204 15:15:04.076678 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fb1088d-906b-41ed-acd5-dc80377be4d4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:04 crc kubenswrapper[4848]: I1204 15:15:04.327970 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" event={"ID":"6fb1088d-906b-41ed-acd5-dc80377be4d4","Type":"ContainerDied","Data":"53eaa9d7fefbad981855ccdf51414990a720d32681db65416c6bb1d3c47f235f"} Dec 04 15:15:04 crc kubenswrapper[4848]: I1204 15:15:04.328012 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53eaa9d7fefbad981855ccdf51414990a720d32681db65416c6bb1d3c47f235f" Dec 04 15:15:04 crc kubenswrapper[4848]: I1204 15:15:04.328064 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414355-66rtt" Dec 04 15:15:04 crc kubenswrapper[4848]: I1204 15:15:04.377794 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm"] Dec 04 15:15:04 crc kubenswrapper[4848]: I1204 15:15:04.388480 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-259bm"] Dec 04 15:15:04 crc kubenswrapper[4848]: I1204 15:15:04.412229 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be9c3abb-a5b2-4e89-8c9d-04b475f528ed" path="/var/lib/kubelet/pods/be9c3abb-a5b2-4e89-8c9d-04b475f528ed/volumes" Dec 04 15:15:07 crc kubenswrapper[4848]: I1204 15:15:07.394324 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:15:07 crc kubenswrapper[4848]: E1204 15:15:07.395233 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.348461 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dp8f4"] Dec 04 15:15:15 crc kubenswrapper[4848]: E1204 15:15:15.349714 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb1088d-906b-41ed-acd5-dc80377be4d4" containerName="collect-profiles" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.349739 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb1088d-906b-41ed-acd5-dc80377be4d4" containerName="collect-profiles" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.350234 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fb1088d-906b-41ed-acd5-dc80377be4d4" containerName="collect-profiles" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.352482 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.408517 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp8f4"] Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.450515 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-utilities\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.450934 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-947d4\" (UniqueName: \"kubernetes.io/projected/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-kube-api-access-947d4\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.451639 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-catalog-content\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.554574 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-catalog-content\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.554712 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-utilities\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.554813 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-947d4\" (UniqueName: \"kubernetes.io/projected/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-kube-api-access-947d4\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.555131 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-catalog-content\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.555256 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-utilities\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.579534 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-947d4\" (UniqueName: \"kubernetes.io/projected/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-kube-api-access-947d4\") pod \"redhat-marketplace-dp8f4\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:15 crc kubenswrapper[4848]: I1204 15:15:15.681003 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:16 crc kubenswrapper[4848]: I1204 15:15:16.151297 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp8f4"] Dec 04 15:15:16 crc kubenswrapper[4848]: I1204 15:15:16.479907 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp8f4" event={"ID":"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238","Type":"ContainerStarted","Data":"907e1f70e64685dd035511baf5ffa707f2eac593dbe4cf273efc9d15aa48bc8c"} Dec 04 15:15:17 crc kubenswrapper[4848]: I1204 15:15:17.491297 4848 generic.go:334] "Generic (PLEG): container finished" podID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerID="f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163" exitCode=0 Dec 04 15:15:17 crc kubenswrapper[4848]: I1204 15:15:17.491346 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp8f4" event={"ID":"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238","Type":"ContainerDied","Data":"f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163"} Dec 04 15:15:19 crc kubenswrapper[4848]: I1204 15:15:19.518162 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp8f4" event={"ID":"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238","Type":"ContainerStarted","Data":"9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88"} Dec 04 15:15:20 crc kubenswrapper[4848]: I1204 15:15:20.531502 4848 generic.go:334] "Generic (PLEG): container finished" podID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerID="9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88" exitCode=0 Dec 04 15:15:20 crc kubenswrapper[4848]: I1204 15:15:20.531568 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp8f4" event={"ID":"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238","Type":"ContainerDied","Data":"9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88"} Dec 04 15:15:21 crc kubenswrapper[4848]: I1204 15:15:21.394018 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:15:21 crc kubenswrapper[4848]: E1204 15:15:21.394935 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:15:21 crc kubenswrapper[4848]: I1204 15:15:21.544159 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp8f4" event={"ID":"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238","Type":"ContainerStarted","Data":"f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a"} Dec 04 15:15:21 crc kubenswrapper[4848]: I1204 15:15:21.570328 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dp8f4" podStartSLOduration=3.135787749 podStartE2EDuration="6.570311465s" podCreationTimestamp="2025-12-04 15:15:15 +0000 UTC" firstStartedPulling="2025-12-04 15:15:17.493815766 +0000 UTC m=+5221.436312294" lastFinishedPulling="2025-12-04 15:15:20.928339482 +0000 UTC m=+5224.870836010" observedRunningTime="2025-12-04 15:15:21.560424821 +0000 UTC m=+5225.502921349" watchObservedRunningTime="2025-12-04 15:15:21.570311465 +0000 UTC m=+5225.512807993" Dec 04 15:15:25 crc kubenswrapper[4848]: I1204 15:15:25.681273 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:25 crc kubenswrapper[4848]: I1204 15:15:25.681874 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:25 crc kubenswrapper[4848]: I1204 15:15:25.726207 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:26 crc kubenswrapper[4848]: I1204 15:15:26.647899 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:26 crc kubenswrapper[4848]: I1204 15:15:26.710274 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp8f4"] Dec 04 15:15:28 crc kubenswrapper[4848]: I1204 15:15:28.611262 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dp8f4" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerName="registry-server" containerID="cri-o://f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a" gracePeriod=2 Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.184857 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.284756 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-utilities\") pod \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.285195 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-947d4\" (UniqueName: \"kubernetes.io/projected/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-kube-api-access-947d4\") pod \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.285338 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-catalog-content\") pod \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\" (UID: \"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238\") " Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.285792 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-utilities" (OuterVolumeSpecName: "utilities") pod "8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" (UID: "8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.286631 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.292136 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-kube-api-access-947d4" (OuterVolumeSpecName: "kube-api-access-947d4") pod "8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" (UID: "8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238"). InnerVolumeSpecName "kube-api-access-947d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.307706 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" (UID: "8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.389372 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-947d4\" (UniqueName: \"kubernetes.io/projected/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-kube-api-access-947d4\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.389617 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.623559 4848 generic.go:334] "Generic (PLEG): container finished" podID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerID="f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a" exitCode=0 Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.623607 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp8f4" event={"ID":"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238","Type":"ContainerDied","Data":"f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a"} Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.623646 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp8f4" event={"ID":"8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238","Type":"ContainerDied","Data":"907e1f70e64685dd035511baf5ffa707f2eac593dbe4cf273efc9d15aa48bc8c"} Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.623651 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dp8f4" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.623663 4848 scope.go:117] "RemoveContainer" containerID="f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.653728 4848 scope.go:117] "RemoveContainer" containerID="9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.676899 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp8f4"] Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.690723 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp8f4"] Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.700047 4848 scope.go:117] "RemoveContainer" containerID="f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.744562 4848 scope.go:117] "RemoveContainer" containerID="f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a" Dec 04 15:15:29 crc kubenswrapper[4848]: E1204 15:15:29.745087 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a\": container with ID starting with f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a not found: ID does not exist" containerID="f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.745356 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a"} err="failed to get container status \"f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a\": rpc error: code = NotFound desc = could not find container \"f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a\": container with ID starting with f2197ef72bca8d1fdf970f796909fb33cc90277df038c9fc6ab1ed885f04b19a not found: ID does not exist" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.745396 4848 scope.go:117] "RemoveContainer" containerID="9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88" Dec 04 15:15:29 crc kubenswrapper[4848]: E1204 15:15:29.745769 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88\": container with ID starting with 9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88 not found: ID does not exist" containerID="9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.745803 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88"} err="failed to get container status \"9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88\": rpc error: code = NotFound desc = could not find container \"9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88\": container with ID starting with 9b412f6504b013439db41251d03aa7134a9d03e5bd06f07deedc474834a41a88 not found: ID does not exist" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.745824 4848 scope.go:117] "RemoveContainer" containerID="f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163" Dec 04 15:15:29 crc kubenswrapper[4848]: E1204 15:15:29.746103 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163\": container with ID starting with f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163 not found: ID does not exist" containerID="f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163" Dec 04 15:15:29 crc kubenswrapper[4848]: I1204 15:15:29.746127 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163"} err="failed to get container status \"f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163\": rpc error: code = NotFound desc = could not find container \"f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163\": container with ID starting with f72fee6420c43d2aa4a36eec3e11f2d60025120d59829aeb952b342e2af5e163 not found: ID does not exist" Dec 04 15:15:30 crc kubenswrapper[4848]: I1204 15:15:30.407649 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" path="/var/lib/kubelet/pods/8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238/volumes" Dec 04 15:15:33 crc kubenswrapper[4848]: I1204 15:15:33.394056 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:15:33 crc kubenswrapper[4848]: E1204 15:15:33.395908 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.816142 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ltc6r"] Dec 04 15:15:34 crc kubenswrapper[4848]: E1204 15:15:34.816886 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerName="extract-content" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.816904 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerName="extract-content" Dec 04 15:15:34 crc kubenswrapper[4848]: E1204 15:15:34.816959 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerName="extract-utilities" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.817532 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerName="extract-utilities" Dec 04 15:15:34 crc kubenswrapper[4848]: E1204 15:15:34.817556 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerName="registry-server" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.817562 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerName="registry-server" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.817794 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="8abd3e5b-b3db-4e0f-b3f4-f5dfecf4d238" containerName="registry-server" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.819561 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.831059 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ltc6r"] Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.921950 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-catalog-content\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.922037 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mbzf\" (UniqueName: \"kubernetes.io/projected/95a057b8-f26f-45d5-8530-f2d985b40368-kube-api-access-7mbzf\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:34 crc kubenswrapper[4848]: I1204 15:15:34.922142 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-utilities\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.024558 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-catalog-content\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.025148 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-catalog-content\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.025433 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mbzf\" (UniqueName: \"kubernetes.io/projected/95a057b8-f26f-45d5-8530-f2d985b40368-kube-api-access-7mbzf\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.025628 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-utilities\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.026286 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-utilities\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.047760 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mbzf\" (UniqueName: \"kubernetes.io/projected/95a057b8-f26f-45d5-8530-f2d985b40368-kube-api-access-7mbzf\") pod \"certified-operators-ltc6r\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.143166 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.662527 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ltc6r"] Dec 04 15:15:35 crc kubenswrapper[4848]: I1204 15:15:35.694507 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltc6r" event={"ID":"95a057b8-f26f-45d5-8530-f2d985b40368","Type":"ContainerStarted","Data":"67008a4ccb5cd8211326bc72233c8ed4dd7f8e75a328b40b6a3cd6fb5f2901ac"} Dec 04 15:15:36 crc kubenswrapper[4848]: I1204 15:15:36.705718 4848 generic.go:334] "Generic (PLEG): container finished" podID="95a057b8-f26f-45d5-8530-f2d985b40368" containerID="0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668" exitCode=0 Dec 04 15:15:36 crc kubenswrapper[4848]: I1204 15:15:36.705792 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltc6r" event={"ID":"95a057b8-f26f-45d5-8530-f2d985b40368","Type":"ContainerDied","Data":"0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668"} Dec 04 15:15:38 crc kubenswrapper[4848]: I1204 15:15:38.730483 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltc6r" event={"ID":"95a057b8-f26f-45d5-8530-f2d985b40368","Type":"ContainerStarted","Data":"41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0"} Dec 04 15:15:39 crc kubenswrapper[4848]: I1204 15:15:39.743023 4848 generic.go:334] "Generic (PLEG): container finished" podID="95a057b8-f26f-45d5-8530-f2d985b40368" containerID="41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0" exitCode=0 Dec 04 15:15:39 crc kubenswrapper[4848]: I1204 15:15:39.743135 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltc6r" event={"ID":"95a057b8-f26f-45d5-8530-f2d985b40368","Type":"ContainerDied","Data":"41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0"} Dec 04 15:15:40 crc kubenswrapper[4848]: I1204 15:15:40.649331 4848 scope.go:117] "RemoveContainer" containerID="ed6f2e078870290acd286ff601629f0742643a3c53bafc05fd47d689b81eb818" Dec 04 15:15:40 crc kubenswrapper[4848]: I1204 15:15:40.763037 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltc6r" event={"ID":"95a057b8-f26f-45d5-8530-f2d985b40368","Type":"ContainerStarted","Data":"f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2"} Dec 04 15:15:40 crc kubenswrapper[4848]: I1204 15:15:40.793135 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ltc6r" podStartSLOduration=3.354698957 podStartE2EDuration="6.79311377s" podCreationTimestamp="2025-12-04 15:15:34 +0000 UTC" firstStartedPulling="2025-12-04 15:15:36.708289635 +0000 UTC m=+5240.650786163" lastFinishedPulling="2025-12-04 15:15:40.146704448 +0000 UTC m=+5244.089200976" observedRunningTime="2025-12-04 15:15:40.788525536 +0000 UTC m=+5244.731022084" watchObservedRunningTime="2025-12-04 15:15:40.79311377 +0000 UTC m=+5244.735610298" Dec 04 15:15:45 crc kubenswrapper[4848]: I1204 15:15:45.143337 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:45 crc kubenswrapper[4848]: I1204 15:15:45.143768 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:45 crc kubenswrapper[4848]: I1204 15:15:45.199789 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:46 crc kubenswrapper[4848]: I1204 15:15:46.205303 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:46 crc kubenswrapper[4848]: I1204 15:15:46.444172 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ltc6r"] Dec 04 15:15:47 crc kubenswrapper[4848]: I1204 15:15:47.393077 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:15:47 crc kubenswrapper[4848]: E1204 15:15:47.393522 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:15:47 crc kubenswrapper[4848]: I1204 15:15:47.841317 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ltc6r" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" containerName="registry-server" containerID="cri-o://f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2" gracePeriod=2 Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.488596 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.591876 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-catalog-content\") pod \"95a057b8-f26f-45d5-8530-f2d985b40368\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.591970 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mbzf\" (UniqueName: \"kubernetes.io/projected/95a057b8-f26f-45d5-8530-f2d985b40368-kube-api-access-7mbzf\") pod \"95a057b8-f26f-45d5-8530-f2d985b40368\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.592243 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-utilities\") pod \"95a057b8-f26f-45d5-8530-f2d985b40368\" (UID: \"95a057b8-f26f-45d5-8530-f2d985b40368\") " Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.593299 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-utilities" (OuterVolumeSpecName: "utilities") pod "95a057b8-f26f-45d5-8530-f2d985b40368" (UID: "95a057b8-f26f-45d5-8530-f2d985b40368"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.601922 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a057b8-f26f-45d5-8530-f2d985b40368-kube-api-access-7mbzf" (OuterVolumeSpecName: "kube-api-access-7mbzf") pod "95a057b8-f26f-45d5-8530-f2d985b40368" (UID: "95a057b8-f26f-45d5-8530-f2d985b40368"). InnerVolumeSpecName "kube-api-access-7mbzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.645492 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95a057b8-f26f-45d5-8530-f2d985b40368" (UID: "95a057b8-f26f-45d5-8530-f2d985b40368"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.696210 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.696260 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mbzf\" (UniqueName: \"kubernetes.io/projected/95a057b8-f26f-45d5-8530-f2d985b40368-kube-api-access-7mbzf\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.696299 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a057b8-f26f-45d5-8530-f2d985b40368-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.855351 4848 generic.go:334] "Generic (PLEG): container finished" podID="95a057b8-f26f-45d5-8530-f2d985b40368" containerID="f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2" exitCode=0 Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.855401 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltc6r" event={"ID":"95a057b8-f26f-45d5-8530-f2d985b40368","Type":"ContainerDied","Data":"f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2"} Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.855413 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltc6r" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.855428 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltc6r" event={"ID":"95a057b8-f26f-45d5-8530-f2d985b40368","Type":"ContainerDied","Data":"67008a4ccb5cd8211326bc72233c8ed4dd7f8e75a328b40b6a3cd6fb5f2901ac"} Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.855443 4848 scope.go:117] "RemoveContainer" containerID="f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.906696 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ltc6r"] Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.910923 4848 scope.go:117] "RemoveContainer" containerID="41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.924181 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ltc6r"] Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.934423 4848 scope.go:117] "RemoveContainer" containerID="0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.991946 4848 scope.go:117] "RemoveContainer" containerID="f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2" Dec 04 15:15:48 crc kubenswrapper[4848]: E1204 15:15:48.992569 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2\": container with ID starting with f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2 not found: ID does not exist" containerID="f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.992608 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2"} err="failed to get container status \"f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2\": rpc error: code = NotFound desc = could not find container \"f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2\": container with ID starting with f65344076851408d92bcddc27246c97ab4d1d261bf8efd269df0f2c6dff56aa2 not found: ID does not exist" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.992637 4848 scope.go:117] "RemoveContainer" containerID="41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0" Dec 04 15:15:48 crc kubenswrapper[4848]: E1204 15:15:48.993009 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0\": container with ID starting with 41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0 not found: ID does not exist" containerID="41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.993063 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0"} err="failed to get container status \"41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0\": rpc error: code = NotFound desc = could not find container \"41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0\": container with ID starting with 41d90c764ae4d025bf1744ddf8e95725f3afb5409a9d14ffdf6eff46f1e3fdf0 not found: ID does not exist" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.993097 4848 scope.go:117] "RemoveContainer" containerID="0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668" Dec 04 15:15:48 crc kubenswrapper[4848]: E1204 15:15:48.993375 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668\": container with ID starting with 0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668 not found: ID does not exist" containerID="0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668" Dec 04 15:15:48 crc kubenswrapper[4848]: I1204 15:15:48.993412 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668"} err="failed to get container status \"0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668\": rpc error: code = NotFound desc = could not find container \"0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668\": container with ID starting with 0d6a071a241a64289a06d266b8d6f513b59b0d71c523627fb27487c1f40ea668 not found: ID does not exist" Dec 04 15:15:50 crc kubenswrapper[4848]: I1204 15:15:50.406679 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" path="/var/lib/kubelet/pods/95a057b8-f26f-45d5-8530-f2d985b40368/volumes" Dec 04 15:16:00 crc kubenswrapper[4848]: I1204 15:16:00.393676 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:16:00 crc kubenswrapper[4848]: E1204 15:16:00.394823 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:16:15 crc kubenswrapper[4848]: I1204 15:16:15.393737 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:16:15 crc kubenswrapper[4848]: E1204 15:16:15.394997 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:16:29 crc kubenswrapper[4848]: I1204 15:16:29.393473 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:16:29 crc kubenswrapper[4848]: E1204 15:16:29.394286 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:16:44 crc kubenswrapper[4848]: I1204 15:16:44.393740 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:16:44 crc kubenswrapper[4848]: E1204 15:16:44.394538 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:16:57 crc kubenswrapper[4848]: I1204 15:16:57.394108 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:16:57 crc kubenswrapper[4848]: E1204 15:16:57.394998 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:17:10 crc kubenswrapper[4848]: I1204 15:17:10.394487 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:17:10 crc kubenswrapper[4848]: E1204 15:17:10.395395 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:17:22 crc kubenswrapper[4848]: I1204 15:17:22.394043 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:17:22 crc kubenswrapper[4848]: E1204 15:17:22.394583 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:17:35 crc kubenswrapper[4848]: I1204 15:17:35.393605 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:17:35 crc kubenswrapper[4848]: E1204 15:17:35.394359 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:17:40 crc kubenswrapper[4848]: I1204 15:17:40.793791 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="c7da8d50-9ec2-4b1d-b373-e65541b02306" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 04 15:17:46 crc kubenswrapper[4848]: I1204 15:17:46.393789 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:17:46 crc kubenswrapper[4848]: E1204 15:17:46.394337 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:17:59 crc kubenswrapper[4848]: I1204 15:17:59.393689 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:17:59 crc kubenswrapper[4848]: E1204 15:17:59.394695 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:18:13 crc kubenswrapper[4848]: I1204 15:18:13.394228 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:18:13 crc kubenswrapper[4848]: E1204 15:18:13.395191 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:18:25 crc kubenswrapper[4848]: I1204 15:18:25.395173 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:18:25 crc kubenswrapper[4848]: E1204 15:18:25.396283 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:18:39 crc kubenswrapper[4848]: I1204 15:18:39.393344 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:18:39 crc kubenswrapper[4848]: E1204 15:18:39.394227 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:18:52 crc kubenswrapper[4848]: I1204 15:18:52.394980 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:18:53 crc kubenswrapper[4848]: I1204 15:18:53.583576 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"2e26022f030c1f2dcc2239563f01b199f010dec6296d6bcdd001cfa9c5ae43cd"} Dec 04 15:21:14 crc kubenswrapper[4848]: I1204 15:21:14.314226 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:21:14 crc kubenswrapper[4848]: I1204 15:21:14.314794 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:21:44 crc kubenswrapper[4848]: I1204 15:21:44.314262 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:21:44 crc kubenswrapper[4848]: I1204 15:21:44.314808 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:22:14 crc kubenswrapper[4848]: I1204 15:22:14.314378 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:22:14 crc kubenswrapper[4848]: I1204 15:22:14.314900 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:22:14 crc kubenswrapper[4848]: I1204 15:22:14.314975 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:22:14 crc kubenswrapper[4848]: I1204 15:22:14.315692 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e26022f030c1f2dcc2239563f01b199f010dec6296d6bcdd001cfa9c5ae43cd"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:22:14 crc kubenswrapper[4848]: I1204 15:22:14.315767 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://2e26022f030c1f2dcc2239563f01b199f010dec6296d6bcdd001cfa9c5ae43cd" gracePeriod=600 Dec 04 15:22:15 crc kubenswrapper[4848]: I1204 15:22:15.109794 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="2e26022f030c1f2dcc2239563f01b199f010dec6296d6bcdd001cfa9c5ae43cd" exitCode=0 Dec 04 15:22:15 crc kubenswrapper[4848]: I1204 15:22:15.109880 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"2e26022f030c1f2dcc2239563f01b199f010dec6296d6bcdd001cfa9c5ae43cd"} Dec 04 15:22:15 crc kubenswrapper[4848]: I1204 15:22:15.110489 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844"} Dec 04 15:22:15 crc kubenswrapper[4848]: I1204 15:22:15.110522 4848 scope.go:117] "RemoveContainer" containerID="fe89dee2d1801fcdf32ee543fb800aff3919496b4205e6cc79f56733169724ba" Dec 04 15:22:47 crc kubenswrapper[4848]: I1204 15:22:47.183571 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-b65c5f4cf-q22f8" podUID="acf82280-f41a-460a-84ab-639c541670fc" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 04 15:24:14 crc kubenswrapper[4848]: I1204 15:24:14.314233 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:24:14 crc kubenswrapper[4848]: I1204 15:24:14.315076 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.532660 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fjkhh"] Dec 04 15:24:36 crc kubenswrapper[4848]: E1204 15:24:36.533803 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" containerName="extract-utilities" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.533829 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" containerName="extract-utilities" Dec 04 15:24:36 crc kubenswrapper[4848]: E1204 15:24:36.533888 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" containerName="registry-server" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.533896 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" containerName="registry-server" Dec 04 15:24:36 crc kubenswrapper[4848]: E1204 15:24:36.533925 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" containerName="extract-content" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.533933 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" containerName="extract-content" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.534225 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a057b8-f26f-45d5-8530-f2d985b40368" containerName="registry-server" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.535940 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.560642 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjkhh"] Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.706179 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-catalog-content\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.706249 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-utilities\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.706474 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns9tm\" (UniqueName: \"kubernetes.io/projected/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-kube-api-access-ns9tm\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.809279 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-catalog-content\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.809352 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-utilities\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.809431 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns9tm\" (UniqueName: \"kubernetes.io/projected/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-kube-api-access-ns9tm\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.810355 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-catalog-content\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.810515 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-utilities\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.830864 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns9tm\" (UniqueName: \"kubernetes.io/projected/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-kube-api-access-ns9tm\") pod \"redhat-operators-fjkhh\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:36 crc kubenswrapper[4848]: I1204 15:24:36.868284 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:37 crc kubenswrapper[4848]: I1204 15:24:37.449587 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjkhh"] Dec 04 15:24:37 crc kubenswrapper[4848]: I1204 15:24:37.704291 4848 generic.go:334] "Generic (PLEG): container finished" podID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerID="3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad" exitCode=0 Dec 04 15:24:37 crc kubenswrapper[4848]: I1204 15:24:37.704397 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjkhh" event={"ID":"f653b2f9-4eec-461d-89ff-8a9b4ad8d782","Type":"ContainerDied","Data":"3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad"} Dec 04 15:24:37 crc kubenswrapper[4848]: I1204 15:24:37.704616 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjkhh" event={"ID":"f653b2f9-4eec-461d-89ff-8a9b4ad8d782","Type":"ContainerStarted","Data":"83e67292cf8739040b6407d2cdce7f09aa5549ee0f088c505d94a857e31ba82b"} Dec 04 15:24:37 crc kubenswrapper[4848]: I1204 15:24:37.707151 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:24:39 crc kubenswrapper[4848]: I1204 15:24:39.727711 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjkhh" event={"ID":"f653b2f9-4eec-461d-89ff-8a9b4ad8d782","Type":"ContainerStarted","Data":"f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702"} Dec 04 15:24:42 crc kubenswrapper[4848]: I1204 15:24:42.763759 4848 generic.go:334] "Generic (PLEG): container finished" podID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerID="f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702" exitCode=0 Dec 04 15:24:42 crc kubenswrapper[4848]: I1204 15:24:42.763861 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjkhh" event={"ID":"f653b2f9-4eec-461d-89ff-8a9b4ad8d782","Type":"ContainerDied","Data":"f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702"} Dec 04 15:24:43 crc kubenswrapper[4848]: I1204 15:24:43.776237 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjkhh" event={"ID":"f653b2f9-4eec-461d-89ff-8a9b4ad8d782","Type":"ContainerStarted","Data":"3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325"} Dec 04 15:24:43 crc kubenswrapper[4848]: I1204 15:24:43.809453 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fjkhh" podStartSLOduration=2.247315049 podStartE2EDuration="7.809428521s" podCreationTimestamp="2025-12-04 15:24:36 +0000 UTC" firstStartedPulling="2025-12-04 15:24:37.706510144 +0000 UTC m=+5781.649006672" lastFinishedPulling="2025-12-04 15:24:43.268623616 +0000 UTC m=+5787.211120144" observedRunningTime="2025-12-04 15:24:43.793229514 +0000 UTC m=+5787.735726042" watchObservedRunningTime="2025-12-04 15:24:43.809428521 +0000 UTC m=+5787.751925049" Dec 04 15:24:44 crc kubenswrapper[4848]: I1204 15:24:44.314522 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:24:44 crc kubenswrapper[4848]: I1204 15:24:44.314604 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:24:46 crc kubenswrapper[4848]: I1204 15:24:46.868994 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:46 crc kubenswrapper[4848]: I1204 15:24:46.869594 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:47 crc kubenswrapper[4848]: I1204 15:24:47.922050 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fjkhh" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="registry-server" probeResult="failure" output=< Dec 04 15:24:47 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:24:47 crc kubenswrapper[4848]: > Dec 04 15:24:56 crc kubenswrapper[4848]: I1204 15:24:56.921993 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:56 crc kubenswrapper[4848]: I1204 15:24:56.976898 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:57 crc kubenswrapper[4848]: I1204 15:24:57.160604 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjkhh"] Dec 04 15:24:58 crc kubenswrapper[4848]: I1204 15:24:58.940207 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fjkhh" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="registry-server" containerID="cri-o://3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325" gracePeriod=2 Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.625091 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.774624 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-utilities\") pod \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.774766 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-catalog-content\") pod \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.774870 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns9tm\" (UniqueName: \"kubernetes.io/projected/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-kube-api-access-ns9tm\") pod \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\" (UID: \"f653b2f9-4eec-461d-89ff-8a9b4ad8d782\") " Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.775469 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-utilities" (OuterVolumeSpecName: "utilities") pod "f653b2f9-4eec-461d-89ff-8a9b4ad8d782" (UID: "f653b2f9-4eec-461d-89ff-8a9b4ad8d782"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.776671 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.782394 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-kube-api-access-ns9tm" (OuterVolumeSpecName: "kube-api-access-ns9tm") pod "f653b2f9-4eec-461d-89ff-8a9b4ad8d782" (UID: "f653b2f9-4eec-461d-89ff-8a9b4ad8d782"). InnerVolumeSpecName "kube-api-access-ns9tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.878778 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns9tm\" (UniqueName: \"kubernetes.io/projected/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-kube-api-access-ns9tm\") on node \"crc\" DevicePath \"\"" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.885037 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f653b2f9-4eec-461d-89ff-8a9b4ad8d782" (UID: "f653b2f9-4eec-461d-89ff-8a9b4ad8d782"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.954742 4848 generic.go:334] "Generic (PLEG): container finished" podID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerID="3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325" exitCode=0 Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.954797 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjkhh" event={"ID":"f653b2f9-4eec-461d-89ff-8a9b4ad8d782","Type":"ContainerDied","Data":"3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325"} Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.954829 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjkhh" event={"ID":"f653b2f9-4eec-461d-89ff-8a9b4ad8d782","Type":"ContainerDied","Data":"83e67292cf8739040b6407d2cdce7f09aa5549ee0f088c505d94a857e31ba82b"} Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.954843 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjkhh" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.954851 4848 scope.go:117] "RemoveContainer" containerID="3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.987154 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f653b2f9-4eec-461d-89ff-8a9b4ad8d782-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.988986 4848 scope.go:117] "RemoveContainer" containerID="f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702" Dec 04 15:24:59 crc kubenswrapper[4848]: I1204 15:24:59.999451 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjkhh"] Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.011780 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fjkhh"] Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.024478 4848 scope.go:117] "RemoveContainer" containerID="3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad" Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.063656 4848 scope.go:117] "RemoveContainer" containerID="3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325" Dec 04 15:25:00 crc kubenswrapper[4848]: E1204 15:25:00.064410 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325\": container with ID starting with 3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325 not found: ID does not exist" containerID="3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325" Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.064602 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325"} err="failed to get container status \"3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325\": rpc error: code = NotFound desc = could not find container \"3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325\": container with ID starting with 3c55577ef4ae57aa568d0eb8870a693017678e094eb9af4d4c88a60f83f98325 not found: ID does not exist" Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.064686 4848 scope.go:117] "RemoveContainer" containerID="f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702" Dec 04 15:25:00 crc kubenswrapper[4848]: E1204 15:25:00.065114 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702\": container with ID starting with f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702 not found: ID does not exist" containerID="f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702" Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.065223 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702"} err="failed to get container status \"f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702\": rpc error: code = NotFound desc = could not find container \"f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702\": container with ID starting with f24ae7502f098fdf848514c86eb3bbefa092044b8eddaaba72a74478af4fd702 not found: ID does not exist" Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.065302 4848 scope.go:117] "RemoveContainer" containerID="3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad" Dec 04 15:25:00 crc kubenswrapper[4848]: E1204 15:25:00.065603 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad\": container with ID starting with 3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad not found: ID does not exist" containerID="3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad" Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.065638 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad"} err="failed to get container status \"3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad\": rpc error: code = NotFound desc = could not find container \"3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad\": container with ID starting with 3d400031b2851d710f1962bd8dea94382d9b1ad1528ad21528c25f23ede273ad not found: ID does not exist" Dec 04 15:25:00 crc kubenswrapper[4848]: I1204 15:25:00.409286 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" path="/var/lib/kubelet/pods/f653b2f9-4eec-461d-89ff-8a9b4ad8d782/volumes" Dec 04 15:25:14 crc kubenswrapper[4848]: I1204 15:25:14.314604 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:25:14 crc kubenswrapper[4848]: I1204 15:25:14.315312 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:25:14 crc kubenswrapper[4848]: I1204 15:25:14.315362 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:25:14 crc kubenswrapper[4848]: I1204 15:25:14.316256 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:25:14 crc kubenswrapper[4848]: I1204 15:25:14.316311 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" gracePeriod=600 Dec 04 15:25:14 crc kubenswrapper[4848]: E1204 15:25:14.438052 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:25:15 crc kubenswrapper[4848]: I1204 15:25:15.162647 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" exitCode=0 Dec 04 15:25:15 crc kubenswrapper[4848]: I1204 15:25:15.162725 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844"} Dec 04 15:25:15 crc kubenswrapper[4848]: I1204 15:25:15.162906 4848 scope.go:117] "RemoveContainer" containerID="2e26022f030c1f2dcc2239563f01b199f010dec6296d6bcdd001cfa9c5ae43cd" Dec 04 15:25:15 crc kubenswrapper[4848]: I1204 15:25:15.163462 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:25:15 crc kubenswrapper[4848]: E1204 15:25:15.163909 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:25:28 crc kubenswrapper[4848]: I1204 15:25:28.393741 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:25:28 crc kubenswrapper[4848]: E1204 15:25:28.394659 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:25:41 crc kubenswrapper[4848]: I1204 15:25:41.393781 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:25:41 crc kubenswrapper[4848]: E1204 15:25:41.394593 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:25:47 crc kubenswrapper[4848]: I1204 15:25:47.912324 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-clg65"] Dec 04 15:25:47 crc kubenswrapper[4848]: E1204 15:25:47.913498 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="extract-utilities" Dec 04 15:25:47 crc kubenswrapper[4848]: I1204 15:25:47.913518 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="extract-utilities" Dec 04 15:25:47 crc kubenswrapper[4848]: E1204 15:25:47.913546 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="registry-server" Dec 04 15:25:47 crc kubenswrapper[4848]: I1204 15:25:47.913555 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="registry-server" Dec 04 15:25:47 crc kubenswrapper[4848]: E1204 15:25:47.913585 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="extract-content" Dec 04 15:25:47 crc kubenswrapper[4848]: I1204 15:25:47.913593 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="extract-content" Dec 04 15:25:47 crc kubenswrapper[4848]: I1204 15:25:47.913893 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="f653b2f9-4eec-461d-89ff-8a9b4ad8d782" containerName="registry-server" Dec 04 15:25:47 crc kubenswrapper[4848]: I1204 15:25:47.916792 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:47 crc kubenswrapper[4848]: I1204 15:25:47.928060 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-clg65"] Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.098422 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-utilities\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.098522 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgptk\" (UniqueName: \"kubernetes.io/projected/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-kube-api-access-tgptk\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.098584 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-catalog-content\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.201488 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-utilities\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.201933 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgptk\" (UniqueName: \"kubernetes.io/projected/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-kube-api-access-tgptk\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.202034 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-catalog-content\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.202262 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-utilities\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.202536 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-catalog-content\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.223458 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgptk\" (UniqueName: \"kubernetes.io/projected/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-kube-api-access-tgptk\") pod \"certified-operators-clg65\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.243845 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:48 crc kubenswrapper[4848]: I1204 15:25:48.746180 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-clg65"] Dec 04 15:25:49 crc kubenswrapper[4848]: I1204 15:25:49.578596 4848 generic.go:334] "Generic (PLEG): container finished" podID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerID="128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c" exitCode=0 Dec 04 15:25:49 crc kubenswrapper[4848]: I1204 15:25:49.580278 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clg65" event={"ID":"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517","Type":"ContainerDied","Data":"128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c"} Dec 04 15:25:49 crc kubenswrapper[4848]: I1204 15:25:49.580404 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clg65" event={"ID":"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517","Type":"ContainerStarted","Data":"9c14cdf12ebdefef6d1b71913c3e25e2747721211939b30159fcdeaba2d7d15d"} Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.091775 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nbgvt"] Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.109523 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.147285 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbgvt"] Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.290350 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-catalog-content\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.290473 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-utilities\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.290660 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqc7z\" (UniqueName: \"kubernetes.io/projected/efedd544-7cfd-457d-894b-991a04bb7e18-kube-api-access-fqc7z\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.392376 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqc7z\" (UniqueName: \"kubernetes.io/projected/efedd544-7cfd-457d-894b-991a04bb7e18-kube-api-access-fqc7z\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.392615 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-catalog-content\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.392660 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-utilities\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.393196 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-utilities\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.393230 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-catalog-content\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.411720 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqc7z\" (UniqueName: \"kubernetes.io/projected/efedd544-7cfd-457d-894b-991a04bb7e18-kube-api-access-fqc7z\") pod \"community-operators-nbgvt\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.444920 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:25:51 crc kubenswrapper[4848]: I1204 15:25:51.646922 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clg65" event={"ID":"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517","Type":"ContainerStarted","Data":"45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f"} Dec 04 15:25:52 crc kubenswrapper[4848]: I1204 15:25:52.094400 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbgvt"] Dec 04 15:25:52 crc kubenswrapper[4848]: W1204 15:25:52.152895 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefedd544_7cfd_457d_894b_991a04bb7e18.slice/crio-98363b6c746e7ab24b53227e362e9a4875a0d3ad0772542bf2ae9fa230155df8 WatchSource:0}: Error finding container 98363b6c746e7ab24b53227e362e9a4875a0d3ad0772542bf2ae9fa230155df8: Status 404 returned error can't find the container with id 98363b6c746e7ab24b53227e362e9a4875a0d3ad0772542bf2ae9fa230155df8 Dec 04 15:25:52 crc kubenswrapper[4848]: I1204 15:25:52.659415 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbgvt" event={"ID":"efedd544-7cfd-457d-894b-991a04bb7e18","Type":"ContainerStarted","Data":"98363b6c746e7ab24b53227e362e9a4875a0d3ad0772542bf2ae9fa230155df8"} Dec 04 15:25:52 crc kubenswrapper[4848]: I1204 15:25:52.661779 4848 generic.go:334] "Generic (PLEG): container finished" podID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerID="45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f" exitCode=0 Dec 04 15:25:52 crc kubenswrapper[4848]: I1204 15:25:52.661809 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clg65" event={"ID":"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517","Type":"ContainerDied","Data":"45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f"} Dec 04 15:25:53 crc kubenswrapper[4848]: I1204 15:25:53.394459 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:25:53 crc kubenswrapper[4848]: E1204 15:25:53.395892 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:25:53 crc kubenswrapper[4848]: I1204 15:25:53.674992 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clg65" event={"ID":"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517","Type":"ContainerStarted","Data":"c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a"} Dec 04 15:25:53 crc kubenswrapper[4848]: I1204 15:25:53.676926 4848 generic.go:334] "Generic (PLEG): container finished" podID="efedd544-7cfd-457d-894b-991a04bb7e18" containerID="74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad" exitCode=0 Dec 04 15:25:53 crc kubenswrapper[4848]: I1204 15:25:53.677001 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbgvt" event={"ID":"efedd544-7cfd-457d-894b-991a04bb7e18","Type":"ContainerDied","Data":"74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad"} Dec 04 15:25:53 crc kubenswrapper[4848]: I1204 15:25:53.711619 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-clg65" podStartSLOduration=3.180629123 podStartE2EDuration="6.711594751s" podCreationTimestamp="2025-12-04 15:25:47 +0000 UTC" firstStartedPulling="2025-12-04 15:25:49.583621766 +0000 UTC m=+5853.526118314" lastFinishedPulling="2025-12-04 15:25:53.114587414 +0000 UTC m=+5857.057083942" observedRunningTime="2025-12-04 15:25:53.695733841 +0000 UTC m=+5857.638230389" watchObservedRunningTime="2025-12-04 15:25:53.711594751 +0000 UTC m=+5857.654091279" Dec 04 15:25:55 crc kubenswrapper[4848]: I1204 15:25:55.704107 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbgvt" event={"ID":"efedd544-7cfd-457d-894b-991a04bb7e18","Type":"ContainerStarted","Data":"6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea"} Dec 04 15:25:56 crc kubenswrapper[4848]: I1204 15:25:56.718619 4848 generic.go:334] "Generic (PLEG): container finished" podID="efedd544-7cfd-457d-894b-991a04bb7e18" containerID="6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea" exitCode=0 Dec 04 15:25:56 crc kubenswrapper[4848]: I1204 15:25:56.718740 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbgvt" event={"ID":"efedd544-7cfd-457d-894b-991a04bb7e18","Type":"ContainerDied","Data":"6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea"} Dec 04 15:25:57 crc kubenswrapper[4848]: I1204 15:25:57.732201 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbgvt" event={"ID":"efedd544-7cfd-457d-894b-991a04bb7e18","Type":"ContainerStarted","Data":"eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b"} Dec 04 15:25:57 crc kubenswrapper[4848]: I1204 15:25:57.762746 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nbgvt" podStartSLOduration=3.304228761 podStartE2EDuration="6.762727257s" podCreationTimestamp="2025-12-04 15:25:51 +0000 UTC" firstStartedPulling="2025-12-04 15:25:53.678432706 +0000 UTC m=+5857.620929234" lastFinishedPulling="2025-12-04 15:25:57.136931212 +0000 UTC m=+5861.079427730" observedRunningTime="2025-12-04 15:25:57.75716631 +0000 UTC m=+5861.699662848" watchObservedRunningTime="2025-12-04 15:25:57.762727257 +0000 UTC m=+5861.705223785" Dec 04 15:25:58 crc kubenswrapper[4848]: I1204 15:25:58.245109 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:58 crc kubenswrapper[4848]: I1204 15:25:58.246728 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.304584 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-clg65" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="registry-server" probeResult="failure" output=< Dec 04 15:25:59 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:25:59 crc kubenswrapper[4848]: > Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.493206 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vqcb2"] Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.496164 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.507589 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vqcb2"] Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.607369 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8jfv\" (UniqueName: \"kubernetes.io/projected/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-kube-api-access-h8jfv\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.607727 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-utilities\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.607840 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-catalog-content\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.710272 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8jfv\" (UniqueName: \"kubernetes.io/projected/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-kube-api-access-h8jfv\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.710384 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-utilities\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.710417 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-catalog-content\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.710969 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-catalog-content\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.711095 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-utilities\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.734331 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8jfv\" (UniqueName: \"kubernetes.io/projected/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-kube-api-access-h8jfv\") pod \"redhat-marketplace-vqcb2\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:25:59 crc kubenswrapper[4848]: I1204 15:25:59.817159 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:26:00 crc kubenswrapper[4848]: I1204 15:26:00.559390 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vqcb2"] Dec 04 15:26:00 crc kubenswrapper[4848]: I1204 15:26:00.764813 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqcb2" event={"ID":"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf","Type":"ContainerStarted","Data":"e42ee5e0499f5cfb07f7bf52b688d36a49835f9a59d648dc3d7f317361a2395a"} Dec 04 15:26:01 crc kubenswrapper[4848]: I1204 15:26:01.445224 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:26:01 crc kubenswrapper[4848]: I1204 15:26:01.445740 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:26:01 crc kubenswrapper[4848]: I1204 15:26:01.778369 4848 generic.go:334] "Generic (PLEG): container finished" podID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerID="baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a" exitCode=0 Dec 04 15:26:01 crc kubenswrapper[4848]: I1204 15:26:01.778422 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqcb2" event={"ID":"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf","Type":"ContainerDied","Data":"baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a"} Dec 04 15:26:02 crc kubenswrapper[4848]: I1204 15:26:02.497932 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-nbgvt" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="registry-server" probeResult="failure" output=< Dec 04 15:26:02 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:26:02 crc kubenswrapper[4848]: > Dec 04 15:26:03 crc kubenswrapper[4848]: I1204 15:26:03.801512 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqcb2" event={"ID":"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf","Type":"ContainerStarted","Data":"90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3"} Dec 04 15:26:04 crc kubenswrapper[4848]: I1204 15:26:04.826392 4848 generic.go:334] "Generic (PLEG): container finished" podID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerID="90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3" exitCode=0 Dec 04 15:26:04 crc kubenswrapper[4848]: I1204 15:26:04.826784 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqcb2" event={"ID":"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf","Type":"ContainerDied","Data":"90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3"} Dec 04 15:26:05 crc kubenswrapper[4848]: I1204 15:26:05.841271 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqcb2" event={"ID":"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf","Type":"ContainerStarted","Data":"5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310"} Dec 04 15:26:05 crc kubenswrapper[4848]: I1204 15:26:05.867526 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vqcb2" podStartSLOduration=3.393514351 podStartE2EDuration="6.867506329s" podCreationTimestamp="2025-12-04 15:25:59 +0000 UTC" firstStartedPulling="2025-12-04 15:26:01.781765592 +0000 UTC m=+5865.724262120" lastFinishedPulling="2025-12-04 15:26:05.25575757 +0000 UTC m=+5869.198254098" observedRunningTime="2025-12-04 15:26:05.858699383 +0000 UTC m=+5869.801195911" watchObservedRunningTime="2025-12-04 15:26:05.867506329 +0000 UTC m=+5869.810002857" Dec 04 15:26:06 crc kubenswrapper[4848]: I1204 15:26:06.402644 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:26:06 crc kubenswrapper[4848]: E1204 15:26:06.403355 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:26:08 crc kubenswrapper[4848]: I1204 15:26:08.293688 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:26:08 crc kubenswrapper[4848]: I1204 15:26:08.342022 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:26:08 crc kubenswrapper[4848]: I1204 15:26:08.680656 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-clg65"] Dec 04 15:26:09 crc kubenswrapper[4848]: I1204 15:26:09.818123 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:26:09 crc kubenswrapper[4848]: I1204 15:26:09.818613 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:26:09 crc kubenswrapper[4848]: I1204 15:26:09.891658 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-clg65" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="registry-server" containerID="cri-o://c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a" gracePeriod=2 Dec 04 15:26:09 crc kubenswrapper[4848]: I1204 15:26:09.965651 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.425841 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.479747 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgptk\" (UniqueName: \"kubernetes.io/projected/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-kube-api-access-tgptk\") pod \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.480075 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-catalog-content\") pod \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.480132 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-utilities\") pod \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\" (UID: \"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517\") " Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.481088 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-utilities" (OuterVolumeSpecName: "utilities") pod "67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" (UID: "67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.487276 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-kube-api-access-tgptk" (OuterVolumeSpecName: "kube-api-access-tgptk") pod "67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" (UID: "67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517"). InnerVolumeSpecName "kube-api-access-tgptk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.541617 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" (UID: "67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.583536 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.583567 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.583576 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgptk\" (UniqueName: \"kubernetes.io/projected/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517-kube-api-access-tgptk\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.917739 4848 generic.go:334] "Generic (PLEG): container finished" podID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerID="c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a" exitCode=0 Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.917794 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clg65" event={"ID":"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517","Type":"ContainerDied","Data":"c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a"} Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.917801 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-clg65" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.917826 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-clg65" event={"ID":"67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517","Type":"ContainerDied","Data":"9c14cdf12ebdefef6d1b71913c3e25e2747721211939b30159fcdeaba2d7d15d"} Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.917846 4848 scope.go:117] "RemoveContainer" containerID="c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.952336 4848 scope.go:117] "RemoveContainer" containerID="45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f" Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.963025 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-clg65"] Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.974501 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-clg65"] Dec 04 15:26:10 crc kubenswrapper[4848]: I1204 15:26:10.982643 4848 scope.go:117] "RemoveContainer" containerID="128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c" Dec 04 15:26:11 crc kubenswrapper[4848]: I1204 15:26:11.036889 4848 scope.go:117] "RemoveContainer" containerID="c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a" Dec 04 15:26:11 crc kubenswrapper[4848]: E1204 15:26:11.037373 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a\": container with ID starting with c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a not found: ID does not exist" containerID="c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a" Dec 04 15:26:11 crc kubenswrapper[4848]: I1204 15:26:11.037404 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a"} err="failed to get container status \"c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a\": rpc error: code = NotFound desc = could not find container \"c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a\": container with ID starting with c29ca1e96991c20a3a50571178a5cc3efabcb9b7030100742cf3e4c891d52d3a not found: ID does not exist" Dec 04 15:26:11 crc kubenswrapper[4848]: I1204 15:26:11.037425 4848 scope.go:117] "RemoveContainer" containerID="45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f" Dec 04 15:26:11 crc kubenswrapper[4848]: E1204 15:26:11.037671 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f\": container with ID starting with 45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f not found: ID does not exist" containerID="45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f" Dec 04 15:26:11 crc kubenswrapper[4848]: I1204 15:26:11.037712 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f"} err="failed to get container status \"45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f\": rpc error: code = NotFound desc = could not find container \"45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f\": container with ID starting with 45d4461d643684fdf07091420ade9c972b16cd847c3579c893946fa4027f8e9f not found: ID does not exist" Dec 04 15:26:11 crc kubenswrapper[4848]: I1204 15:26:11.037741 4848 scope.go:117] "RemoveContainer" containerID="128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c" Dec 04 15:26:11 crc kubenswrapper[4848]: E1204 15:26:11.038034 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c\": container with ID starting with 128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c not found: ID does not exist" containerID="128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c" Dec 04 15:26:11 crc kubenswrapper[4848]: I1204 15:26:11.038073 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c"} err="failed to get container status \"128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c\": rpc error: code = NotFound desc = could not find container \"128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c\": container with ID starting with 128df323849b45141372e61b9f424ebbe1eec887e88676f1d165d8ccdfebee6c not found: ID does not exist" Dec 04 15:26:11 crc kubenswrapper[4848]: I1204 15:26:11.495318 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:26:11 crc kubenswrapper[4848]: I1204 15:26:11.556507 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:26:12 crc kubenswrapper[4848]: I1204 15:26:12.406928 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" path="/var/lib/kubelet/pods/67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517/volumes" Dec 04 15:26:13 crc kubenswrapper[4848]: I1204 15:26:13.880872 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nbgvt"] Dec 04 15:26:13 crc kubenswrapper[4848]: I1204 15:26:13.881395 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nbgvt" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="registry-server" containerID="cri-o://eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b" gracePeriod=2 Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.394669 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.482013 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-catalog-content\") pod \"efedd544-7cfd-457d-894b-991a04bb7e18\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.482596 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-utilities\") pod \"efedd544-7cfd-457d-894b-991a04bb7e18\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.482939 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqc7z\" (UniqueName: \"kubernetes.io/projected/efedd544-7cfd-457d-894b-991a04bb7e18-kube-api-access-fqc7z\") pod \"efedd544-7cfd-457d-894b-991a04bb7e18\" (UID: \"efedd544-7cfd-457d-894b-991a04bb7e18\") " Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.483366 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-utilities" (OuterVolumeSpecName: "utilities") pod "efedd544-7cfd-457d-894b-991a04bb7e18" (UID: "efedd544-7cfd-457d-894b-991a04bb7e18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.484218 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.491502 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efedd544-7cfd-457d-894b-991a04bb7e18-kube-api-access-fqc7z" (OuterVolumeSpecName: "kube-api-access-fqc7z") pod "efedd544-7cfd-457d-894b-991a04bb7e18" (UID: "efedd544-7cfd-457d-894b-991a04bb7e18"). InnerVolumeSpecName "kube-api-access-fqc7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.537891 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efedd544-7cfd-457d-894b-991a04bb7e18" (UID: "efedd544-7cfd-457d-894b-991a04bb7e18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.586135 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efedd544-7cfd-457d-894b-991a04bb7e18-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.586172 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqc7z\" (UniqueName: \"kubernetes.io/projected/efedd544-7cfd-457d-894b-991a04bb7e18-kube-api-access-fqc7z\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.969386 4848 generic.go:334] "Generic (PLEG): container finished" podID="efedd544-7cfd-457d-894b-991a04bb7e18" containerID="eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b" exitCode=0 Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.969432 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbgvt" event={"ID":"efedd544-7cfd-457d-894b-991a04bb7e18","Type":"ContainerDied","Data":"eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b"} Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.969459 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbgvt" event={"ID":"efedd544-7cfd-457d-894b-991a04bb7e18","Type":"ContainerDied","Data":"98363b6c746e7ab24b53227e362e9a4875a0d3ad0772542bf2ae9fa230155df8"} Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.969475 4848 scope.go:117] "RemoveContainer" containerID="eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b" Dec 04 15:26:14 crc kubenswrapper[4848]: I1204 15:26:14.969485 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbgvt" Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.001590 4848 scope.go:117] "RemoveContainer" containerID="6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea" Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.027805 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nbgvt"] Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.041322 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nbgvt"] Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.042916 4848 scope.go:117] "RemoveContainer" containerID="74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad" Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.085199 4848 scope.go:117] "RemoveContainer" containerID="eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b" Dec 04 15:26:15 crc kubenswrapper[4848]: E1204 15:26:15.086446 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b\": container with ID starting with eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b not found: ID does not exist" containerID="eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b" Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.086500 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b"} err="failed to get container status \"eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b\": rpc error: code = NotFound desc = could not find container \"eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b\": container with ID starting with eba9849aab4a0d8072832727589e94968857de1d4a23d45a42429f56f4e7ff0b not found: ID does not exist" Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.086531 4848 scope.go:117] "RemoveContainer" containerID="6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea" Dec 04 15:26:15 crc kubenswrapper[4848]: E1204 15:26:15.086852 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea\": container with ID starting with 6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea not found: ID does not exist" containerID="6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea" Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.086883 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea"} err="failed to get container status \"6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea\": rpc error: code = NotFound desc = could not find container \"6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea\": container with ID starting with 6148816dab832e3f3c85dd85e3cc2a43ebc3ec2c78e7ffe840b6f0c0506246ea not found: ID does not exist" Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.086901 4848 scope.go:117] "RemoveContainer" containerID="74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad" Dec 04 15:26:15 crc kubenswrapper[4848]: E1204 15:26:15.087196 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad\": container with ID starting with 74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad not found: ID does not exist" containerID="74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad" Dec 04 15:26:15 crc kubenswrapper[4848]: I1204 15:26:15.087229 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad"} err="failed to get container status \"74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad\": rpc error: code = NotFound desc = could not find container \"74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad\": container with ID starting with 74451dccf962c0fc0b1132840822c43959f5000efe3636e67cb8300bdeb65dad not found: ID does not exist" Dec 04 15:26:16 crc kubenswrapper[4848]: I1204 15:26:16.410639 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" path="/var/lib/kubelet/pods/efedd544-7cfd-457d-894b-991a04bb7e18/volumes" Dec 04 15:26:18 crc kubenswrapper[4848]: I1204 15:26:18.394027 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:26:18 crc kubenswrapper[4848]: E1204 15:26:18.394750 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:26:19 crc kubenswrapper[4848]: I1204 15:26:19.869675 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:26:19 crc kubenswrapper[4848]: I1204 15:26:19.929304 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vqcb2"] Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.023708 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vqcb2" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerName="registry-server" containerID="cri-o://5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310" gracePeriod=2 Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.504626 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.630346 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-utilities\") pod \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.630446 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8jfv\" (UniqueName: \"kubernetes.io/projected/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-kube-api-access-h8jfv\") pod \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.630600 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-catalog-content\") pod \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\" (UID: \"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf\") " Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.631888 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-utilities" (OuterVolumeSpecName: "utilities") pod "0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" (UID: "0b5cbaea-e0e9-4a7f-8d82-914035cb11bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.637142 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-kube-api-access-h8jfv" (OuterVolumeSpecName: "kube-api-access-h8jfv") pod "0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" (UID: "0b5cbaea-e0e9-4a7f-8d82-914035cb11bf"). InnerVolumeSpecName "kube-api-access-h8jfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.655728 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" (UID: "0b5cbaea-e0e9-4a7f-8d82-914035cb11bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.733797 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.733840 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8jfv\" (UniqueName: \"kubernetes.io/projected/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-kube-api-access-h8jfv\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:20 crc kubenswrapper[4848]: I1204 15:26:20.733852 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.035436 4848 generic.go:334] "Generic (PLEG): container finished" podID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerID="5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310" exitCode=0 Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.035498 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqcb2" event={"ID":"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf","Type":"ContainerDied","Data":"5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310"} Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.035514 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vqcb2" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.035575 4848 scope.go:117] "RemoveContainer" containerID="5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.035560 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vqcb2" event={"ID":"0b5cbaea-e0e9-4a7f-8d82-914035cb11bf","Type":"ContainerDied","Data":"e42ee5e0499f5cfb07f7bf52b688d36a49835f9a59d648dc3d7f317361a2395a"} Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.063142 4848 scope.go:117] "RemoveContainer" containerID="90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.069267 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vqcb2"] Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.081883 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vqcb2"] Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.095204 4848 scope.go:117] "RemoveContainer" containerID="baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.138894 4848 scope.go:117] "RemoveContainer" containerID="5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310" Dec 04 15:26:21 crc kubenswrapper[4848]: E1204 15:26:21.139543 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310\": container with ID starting with 5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310 not found: ID does not exist" containerID="5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.139572 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310"} err="failed to get container status \"5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310\": rpc error: code = NotFound desc = could not find container \"5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310\": container with ID starting with 5a724089ee0e7caa081d12518e145e6aeee5583658d2c91644e3fafd8ea9a310 not found: ID does not exist" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.139592 4848 scope.go:117] "RemoveContainer" containerID="90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3" Dec 04 15:26:21 crc kubenswrapper[4848]: E1204 15:26:21.139795 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3\": container with ID starting with 90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3 not found: ID does not exist" containerID="90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.139815 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3"} err="failed to get container status \"90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3\": rpc error: code = NotFound desc = could not find container \"90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3\": container with ID starting with 90ef8d9ce78865f42e529f019099e66dbc570a039cd0bfd8d5755b651e38aea3 not found: ID does not exist" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.139828 4848 scope.go:117] "RemoveContainer" containerID="baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a" Dec 04 15:26:21 crc kubenswrapper[4848]: E1204 15:26:21.140051 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a\": container with ID starting with baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a not found: ID does not exist" containerID="baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a" Dec 04 15:26:21 crc kubenswrapper[4848]: I1204 15:26:21.140079 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a"} err="failed to get container status \"baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a\": rpc error: code = NotFound desc = could not find container \"baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a\": container with ID starting with baac214d0781620393fab4e6e631857d1e18bba790a748da992bf1647868ac4a not found: ID does not exist" Dec 04 15:26:22 crc kubenswrapper[4848]: I1204 15:26:22.408948 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" path="/var/lib/kubelet/pods/0b5cbaea-e0e9-4a7f-8d82-914035cb11bf/volumes" Dec 04 15:26:32 crc kubenswrapper[4848]: I1204 15:26:32.394078 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:26:32 crc kubenswrapper[4848]: E1204 15:26:32.394921 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:26:41 crc kubenswrapper[4848]: I1204 15:26:41.903265 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-zjnwv" podUID="55609dec-d4b0-421b-b2c9-021978925d25" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 15:26:45 crc kubenswrapper[4848]: I1204 15:26:45.394442 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:26:45 crc kubenswrapper[4848]: E1204 15:26:45.396156 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:26:59 crc kubenswrapper[4848]: I1204 15:26:59.394809 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:26:59 crc kubenswrapper[4848]: E1204 15:26:59.395512 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:27:12 crc kubenswrapper[4848]: I1204 15:27:12.393331 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:27:12 crc kubenswrapper[4848]: E1204 15:27:12.394093 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:27:23 crc kubenswrapper[4848]: I1204 15:27:23.393496 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:27:23 crc kubenswrapper[4848]: E1204 15:27:23.394253 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:27:34 crc kubenswrapper[4848]: I1204 15:27:34.393329 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:27:34 crc kubenswrapper[4848]: E1204 15:27:34.395011 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:27:47 crc kubenswrapper[4848]: I1204 15:27:47.394153 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:27:47 crc kubenswrapper[4848]: E1204 15:27:47.396512 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:27:58 crc kubenswrapper[4848]: I1204 15:27:58.394282 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:27:58 crc kubenswrapper[4848]: E1204 15:27:58.395267 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:28:10 crc kubenswrapper[4848]: I1204 15:28:10.393850 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:28:10 crc kubenswrapper[4848]: E1204 15:28:10.394719 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:28:23 crc kubenswrapper[4848]: I1204 15:28:23.392930 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:28:23 crc kubenswrapper[4848]: E1204 15:28:23.393506 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:28:36 crc kubenswrapper[4848]: I1204 15:28:36.406019 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:28:36 crc kubenswrapper[4848]: E1204 15:28:36.407032 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:28:47 crc kubenswrapper[4848]: I1204 15:28:47.393993 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:28:47 crc kubenswrapper[4848]: E1204 15:28:47.394725 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:29:00 crc kubenswrapper[4848]: I1204 15:29:00.394859 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:29:00 crc kubenswrapper[4848]: E1204 15:29:00.395733 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:29:12 crc kubenswrapper[4848]: I1204 15:29:12.394489 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:29:12 crc kubenswrapper[4848]: E1204 15:29:12.395508 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:29:27 crc kubenswrapper[4848]: I1204 15:29:27.393518 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:29:27 crc kubenswrapper[4848]: E1204 15:29:27.394463 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:29:37 crc kubenswrapper[4848]: I1204 15:29:37.958722 4848 generic.go:334] "Generic (PLEG): container finished" podID="9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" containerID="290ba3831b53925b3626369bf155fb33b9789c8116171885778ba9c197cae4c3" exitCode=0 Dec 04 15:29:37 crc kubenswrapper[4848]: I1204 15:29:37.958817 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a","Type":"ContainerDied","Data":"290ba3831b53925b3626369bf155fb33b9789c8116171885778ba9c197cae4c3"} Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.369339 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.394318 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:29:39 crc kubenswrapper[4848]: E1204 15:29:39.394729 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467180 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-temporary\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467306 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config-secret\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467361 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ssh-key\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467522 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-config-data\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467620 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ca-certs\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467659 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2thsw\" (UniqueName: \"kubernetes.io/projected/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-kube-api-access-2thsw\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467688 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467737 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.467834 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-workdir\") pod \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\" (UID: \"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a\") " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.468392 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.469645 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-config-data" (OuterVolumeSpecName: "config-data") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.471880 4848 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.471919 4848 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.475406 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-kube-api-access-2thsw" (OuterVolumeSpecName: "kube-api-access-2thsw") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "kube-api-access-2thsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.475848 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.478052 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.505871 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.513396 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.516250 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.533940 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" (UID: "9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.573652 4848 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.573686 4848 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.573696 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2thsw\" (UniqueName: \"kubernetes.io/projected/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-kube-api-access-2thsw\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.573708 4848 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.574051 4848 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.574075 4848 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.574088 4848 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.601496 4848 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 04 15:29:39 crc kubenswrapper[4848]: I1204 15:29:39.676328 4848 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 04 15:29:40 crc kubenswrapper[4848]: I1204 15:29:39.983306 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a","Type":"ContainerDied","Data":"046bd5f1ac0c0f08de9648b6733b739264c1de352b8a16ec2fe6994ddda07bc7"} Dec 04 15:29:40 crc kubenswrapper[4848]: I1204 15:29:39.983338 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 15:29:40 crc kubenswrapper[4848]: I1204 15:29:39.983345 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="046bd5f1ac0c0f08de9648b6733b739264c1de352b8a16ec2fe6994ddda07bc7" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.890236 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891570 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="extract-content" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891595 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="extract-content" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891611 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerName="extract-utilities" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891622 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerName="extract-utilities" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891648 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerName="extract-content" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891659 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerName="extract-content" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891677 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891688 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891709 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" containerName="tempest-tests-tempest-tests-runner" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891721 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" containerName="tempest-tests-tempest-tests-runner" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891742 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891752 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891797 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="extract-utilities" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891817 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="extract-utilities" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891842 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="extract-utilities" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891851 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="extract-utilities" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891863 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891871 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: E1204 15:29:44.891893 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="extract-content" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.891907 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="extract-content" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.892221 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="efedd544-7cfd-457d-894b-991a04bb7e18" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.892271 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a" containerName="tempest-tests-tempest-tests-runner" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.892285 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b5cbaea-e0e9-4a7f-8d82-914035cb11bf" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.892305 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="67fa9ce1-7ec5-40f3-a7a3-b2beb68a0517" containerName="registry-server" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.893519 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.897347 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-kfvbp" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.904686 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.991150 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjwlx\" (UniqueName: \"kubernetes.io/projected/3d34d919-565d-49be-8170-1bec9816c2c9-kube-api-access-zjwlx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d34d919-565d-49be-8170-1bec9816c2c9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:44 crc kubenswrapper[4848]: I1204 15:29:44.991878 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d34d919-565d-49be-8170-1bec9816c2c9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:45 crc kubenswrapper[4848]: I1204 15:29:45.094100 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d34d919-565d-49be-8170-1bec9816c2c9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:45 crc kubenswrapper[4848]: I1204 15:29:45.094243 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjwlx\" (UniqueName: \"kubernetes.io/projected/3d34d919-565d-49be-8170-1bec9816c2c9-kube-api-access-zjwlx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d34d919-565d-49be-8170-1bec9816c2c9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:45 crc kubenswrapper[4848]: I1204 15:29:45.096261 4848 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d34d919-565d-49be-8170-1bec9816c2c9\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:45 crc kubenswrapper[4848]: I1204 15:29:45.115027 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjwlx\" (UniqueName: \"kubernetes.io/projected/3d34d919-565d-49be-8170-1bec9816c2c9-kube-api-access-zjwlx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d34d919-565d-49be-8170-1bec9816c2c9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:45 crc kubenswrapper[4848]: I1204 15:29:45.131097 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3d34d919-565d-49be-8170-1bec9816c2c9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:45 crc kubenswrapper[4848]: I1204 15:29:45.221903 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 15:29:45 crc kubenswrapper[4848]: I1204 15:29:45.770905 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 15:29:45 crc kubenswrapper[4848]: I1204 15:29:45.781035 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:29:46 crc kubenswrapper[4848]: I1204 15:29:46.045993 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"3d34d919-565d-49be-8170-1bec9816c2c9","Type":"ContainerStarted","Data":"ac005b39084f2a6fc8e1fdd78c83e7b11918430ec9de0652d1a08ed8a447328d"} Dec 04 15:29:47 crc kubenswrapper[4848]: I1204 15:29:47.060369 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"3d34d919-565d-49be-8170-1bec9816c2c9","Type":"ContainerStarted","Data":"19d45ee22d8139e9132a75c8403285fb04234a01b17d29e1c5657363ac1bd9b6"} Dec 04 15:29:47 crc kubenswrapper[4848]: I1204 15:29:47.085644 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.156686655 podStartE2EDuration="3.08562018s" podCreationTimestamp="2025-12-04 15:29:44 +0000 UTC" firstStartedPulling="2025-12-04 15:29:45.780783264 +0000 UTC m=+6089.723279792" lastFinishedPulling="2025-12-04 15:29:46.709716789 +0000 UTC m=+6090.652213317" observedRunningTime="2025-12-04 15:29:47.075770319 +0000 UTC m=+6091.018266847" watchObservedRunningTime="2025-12-04 15:29:47.08562018 +0000 UTC m=+6091.028116708" Dec 04 15:29:54 crc kubenswrapper[4848]: I1204 15:29:54.394192 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:29:54 crc kubenswrapper[4848]: E1204 15:29:54.395133 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.156497 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp"] Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.159020 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.163017 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.163359 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.175112 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp"] Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.258202 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b438cac7-03a9-4b1e-ab4e-4848a2679da0-secret-volume\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.258493 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b438cac7-03a9-4b1e-ab4e-4848a2679da0-config-volume\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.258580 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcl9h\" (UniqueName: \"kubernetes.io/projected/b438cac7-03a9-4b1e-ab4e-4848a2679da0-kube-api-access-rcl9h\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.361116 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b438cac7-03a9-4b1e-ab4e-4848a2679da0-secret-volume\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.361573 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b438cac7-03a9-4b1e-ab4e-4848a2679da0-config-volume\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.361621 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcl9h\" (UniqueName: \"kubernetes.io/projected/b438cac7-03a9-4b1e-ab4e-4848a2679da0-kube-api-access-rcl9h\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.362575 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b438cac7-03a9-4b1e-ab4e-4848a2679da0-config-volume\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.367057 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b438cac7-03a9-4b1e-ab4e-4848a2679da0-secret-volume\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.379034 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcl9h\" (UniqueName: \"kubernetes.io/projected/b438cac7-03a9-4b1e-ab4e-4848a2679da0-kube-api-access-rcl9h\") pod \"collect-profiles-29414370-7dhtp\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.483761 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:00 crc kubenswrapper[4848]: I1204 15:30:00.962352 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp"] Dec 04 15:30:00 crc kubenswrapper[4848]: W1204 15:30:00.967327 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb438cac7_03a9_4b1e_ab4e_4848a2679da0.slice/crio-34fbccdd9806fc61d576753d7fa1431d0d43fa9cf7ae169f09fdec9ccbe03198 WatchSource:0}: Error finding container 34fbccdd9806fc61d576753d7fa1431d0d43fa9cf7ae169f09fdec9ccbe03198: Status 404 returned error can't find the container with id 34fbccdd9806fc61d576753d7fa1431d0d43fa9cf7ae169f09fdec9ccbe03198 Dec 04 15:30:01 crc kubenswrapper[4848]: I1204 15:30:01.224684 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" event={"ID":"b438cac7-03a9-4b1e-ab4e-4848a2679da0","Type":"ContainerStarted","Data":"dedb3dd41b3097efe4f9457d4b30ba45f4fcd927d4b51b2f0d279db39b111998"} Dec 04 15:30:01 crc kubenswrapper[4848]: I1204 15:30:01.225090 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" event={"ID":"b438cac7-03a9-4b1e-ab4e-4848a2679da0","Type":"ContainerStarted","Data":"34fbccdd9806fc61d576753d7fa1431d0d43fa9cf7ae169f09fdec9ccbe03198"} Dec 04 15:30:01 crc kubenswrapper[4848]: I1204 15:30:01.252141 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" podStartSLOduration=1.2521171500000001 podStartE2EDuration="1.25211715s" podCreationTimestamp="2025-12-04 15:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 15:30:01.240196409 +0000 UTC m=+6105.182692947" watchObservedRunningTime="2025-12-04 15:30:01.25211715 +0000 UTC m=+6105.194613678" Dec 04 15:30:02 crc kubenswrapper[4848]: I1204 15:30:02.238592 4848 generic.go:334] "Generic (PLEG): container finished" podID="b438cac7-03a9-4b1e-ab4e-4848a2679da0" containerID="dedb3dd41b3097efe4f9457d4b30ba45f4fcd927d4b51b2f0d279db39b111998" exitCode=0 Dec 04 15:30:02 crc kubenswrapper[4848]: I1204 15:30:02.238683 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" event={"ID":"b438cac7-03a9-4b1e-ab4e-4848a2679da0","Type":"ContainerDied","Data":"dedb3dd41b3097efe4f9457d4b30ba45f4fcd927d4b51b2f0d279db39b111998"} Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.660443 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.748586 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b438cac7-03a9-4b1e-ab4e-4848a2679da0-config-volume\") pod \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.748630 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcl9h\" (UniqueName: \"kubernetes.io/projected/b438cac7-03a9-4b1e-ab4e-4848a2679da0-kube-api-access-rcl9h\") pod \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.748665 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b438cac7-03a9-4b1e-ab4e-4848a2679da0-secret-volume\") pod \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\" (UID: \"b438cac7-03a9-4b1e-ab4e-4848a2679da0\") " Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.749359 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b438cac7-03a9-4b1e-ab4e-4848a2679da0-config-volume" (OuterVolumeSpecName: "config-volume") pod "b438cac7-03a9-4b1e-ab4e-4848a2679da0" (UID: "b438cac7-03a9-4b1e-ab4e-4848a2679da0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.754810 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b438cac7-03a9-4b1e-ab4e-4848a2679da0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b438cac7-03a9-4b1e-ab4e-4848a2679da0" (UID: "b438cac7-03a9-4b1e-ab4e-4848a2679da0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.755455 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b438cac7-03a9-4b1e-ab4e-4848a2679da0-kube-api-access-rcl9h" (OuterVolumeSpecName: "kube-api-access-rcl9h") pod "b438cac7-03a9-4b1e-ab4e-4848a2679da0" (UID: "b438cac7-03a9-4b1e-ab4e-4848a2679da0"). InnerVolumeSpecName "kube-api-access-rcl9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.851370 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b438cac7-03a9-4b1e-ab4e-4848a2679da0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.851407 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcl9h\" (UniqueName: \"kubernetes.io/projected/b438cac7-03a9-4b1e-ab4e-4848a2679da0-kube-api-access-rcl9h\") on node \"crc\" DevicePath \"\"" Dec 04 15:30:03 crc kubenswrapper[4848]: I1204 15:30:03.851419 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b438cac7-03a9-4b1e-ab4e-4848a2679da0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:30:04 crc kubenswrapper[4848]: I1204 15:30:04.262676 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" event={"ID":"b438cac7-03a9-4b1e-ab4e-4848a2679da0","Type":"ContainerDied","Data":"34fbccdd9806fc61d576753d7fa1431d0d43fa9cf7ae169f09fdec9ccbe03198"} Dec 04 15:30:04 crc kubenswrapper[4848]: I1204 15:30:04.262720 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34fbccdd9806fc61d576753d7fa1431d0d43fa9cf7ae169f09fdec9ccbe03198" Dec 04 15:30:04 crc kubenswrapper[4848]: I1204 15:30:04.263086 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414370-7dhtp" Dec 04 15:30:04 crc kubenswrapper[4848]: I1204 15:30:04.339447 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8"] Dec 04 15:30:04 crc kubenswrapper[4848]: I1204 15:30:04.348548 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414325-nj2m8"] Dec 04 15:30:04 crc kubenswrapper[4848]: I1204 15:30:04.408238 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92487436-fbca-4b93-b6ab-1fa9d5686522" path="/var/lib/kubelet/pods/92487436-fbca-4b93-b6ab-1fa9d5686522/volumes" Dec 04 15:30:09 crc kubenswrapper[4848]: I1204 15:30:09.393665 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:30:09 crc kubenswrapper[4848]: E1204 15:30:09.394643 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.850003 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m2chb/must-gather-pjmd4"] Dec 04 15:30:17 crc kubenswrapper[4848]: E1204 15:30:17.851149 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b438cac7-03a9-4b1e-ab4e-4848a2679da0" containerName="collect-profiles" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.851172 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b438cac7-03a9-4b1e-ab4e-4848a2679da0" containerName="collect-profiles" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.851510 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b438cac7-03a9-4b1e-ab4e-4848a2679da0" containerName="collect-profiles" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.853360 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.855787 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-m2chb"/"default-dockercfg-tn69p" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.856017 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m2chb"/"kube-root-ca.crt" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.856752 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m2chb"/"openshift-service-ca.crt" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.866803 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m2chb/must-gather-pjmd4"] Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.901504 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vplrz\" (UniqueName: \"kubernetes.io/projected/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-kube-api-access-vplrz\") pod \"must-gather-pjmd4\" (UID: \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\") " pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:30:17 crc kubenswrapper[4848]: I1204 15:30:17.901721 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-must-gather-output\") pod \"must-gather-pjmd4\" (UID: \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\") " pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:30:18 crc kubenswrapper[4848]: I1204 15:30:18.003389 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-must-gather-output\") pod \"must-gather-pjmd4\" (UID: \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\") " pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:30:18 crc kubenswrapper[4848]: I1204 15:30:18.003510 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vplrz\" (UniqueName: \"kubernetes.io/projected/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-kube-api-access-vplrz\") pod \"must-gather-pjmd4\" (UID: \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\") " pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:30:18 crc kubenswrapper[4848]: I1204 15:30:18.003879 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-must-gather-output\") pod \"must-gather-pjmd4\" (UID: \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\") " pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:30:18 crc kubenswrapper[4848]: I1204 15:30:18.032530 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vplrz\" (UniqueName: \"kubernetes.io/projected/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-kube-api-access-vplrz\") pod \"must-gather-pjmd4\" (UID: \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\") " pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:30:18 crc kubenswrapper[4848]: I1204 15:30:18.174098 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:30:18 crc kubenswrapper[4848]: I1204 15:30:18.779907 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m2chb/must-gather-pjmd4"] Dec 04 15:30:19 crc kubenswrapper[4848]: I1204 15:30:19.438493 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/must-gather-pjmd4" event={"ID":"38bd2da4-bb47-492f-9bd3-fd64625a0ceb","Type":"ContainerStarted","Data":"50136b285e74f46a23c930e33bc7b75cb6efff84e62c2cc37f96a1be44c9ee67"} Dec 04 15:30:23 crc kubenswrapper[4848]: I1204 15:30:23.483286 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/must-gather-pjmd4" event={"ID":"38bd2da4-bb47-492f-9bd3-fd64625a0ceb","Type":"ContainerStarted","Data":"edb90f767b4c56392ebbaa205266253d073cb5cdfac6b23d7036a3c08a10b1bc"} Dec 04 15:30:23 crc kubenswrapper[4848]: I1204 15:30:23.483801 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/must-gather-pjmd4" event={"ID":"38bd2da4-bb47-492f-9bd3-fd64625a0ceb","Type":"ContainerStarted","Data":"579b3ae0082fabe61435e5fcf9bc6d06ca43a91cea8ea591c17c04042fa131d4"} Dec 04 15:30:23 crc kubenswrapper[4848]: I1204 15:30:23.500275 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m2chb/must-gather-pjmd4" podStartSLOduration=2.758342792 podStartE2EDuration="6.500257853s" podCreationTimestamp="2025-12-04 15:30:17 +0000 UTC" firstStartedPulling="2025-12-04 15:30:18.77688475 +0000 UTC m=+6122.719381278" lastFinishedPulling="2025-12-04 15:30:22.518799791 +0000 UTC m=+6126.461296339" observedRunningTime="2025-12-04 15:30:23.497015554 +0000 UTC m=+6127.439512082" watchObservedRunningTime="2025-12-04 15:30:23.500257853 +0000 UTC m=+6127.442754381" Dec 04 15:30:24 crc kubenswrapper[4848]: I1204 15:30:24.394150 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:30:25 crc kubenswrapper[4848]: I1204 15:30:25.518447 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"6e8b1d4f65c5c25aa19d0b9bebebf377fde86d7d15e1e29fa678a12b8a93b1e9"} Dec 04 15:30:27 crc kubenswrapper[4848]: E1204 15:30:27.354946 4848 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.5:54628->38.102.83.5:35999: write tcp 38.102.83.5:54628->38.102.83.5:35999: write: broken pipe Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.317780 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m2chb/crc-debug-lxthz"] Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.319692 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.454405 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sc9b\" (UniqueName: \"kubernetes.io/projected/8abb634d-2abe-444b-8ab5-04ab450086b7-kube-api-access-2sc9b\") pod \"crc-debug-lxthz\" (UID: \"8abb634d-2abe-444b-8ab5-04ab450086b7\") " pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.455052 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8abb634d-2abe-444b-8ab5-04ab450086b7-host\") pod \"crc-debug-lxthz\" (UID: \"8abb634d-2abe-444b-8ab5-04ab450086b7\") " pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.557203 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sc9b\" (UniqueName: \"kubernetes.io/projected/8abb634d-2abe-444b-8ab5-04ab450086b7-kube-api-access-2sc9b\") pod \"crc-debug-lxthz\" (UID: \"8abb634d-2abe-444b-8ab5-04ab450086b7\") " pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.557486 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8abb634d-2abe-444b-8ab5-04ab450086b7-host\") pod \"crc-debug-lxthz\" (UID: \"8abb634d-2abe-444b-8ab5-04ab450086b7\") " pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.558002 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8abb634d-2abe-444b-8ab5-04ab450086b7-host\") pod \"crc-debug-lxthz\" (UID: \"8abb634d-2abe-444b-8ab5-04ab450086b7\") " pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.577751 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sc9b\" (UniqueName: \"kubernetes.io/projected/8abb634d-2abe-444b-8ab5-04ab450086b7-kube-api-access-2sc9b\") pod \"crc-debug-lxthz\" (UID: \"8abb634d-2abe-444b-8ab5-04ab450086b7\") " pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:30:28 crc kubenswrapper[4848]: I1204 15:30:28.640971 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:30:29 crc kubenswrapper[4848]: I1204 15:30:29.570119 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/crc-debug-lxthz" event={"ID":"8abb634d-2abe-444b-8ab5-04ab450086b7","Type":"ContainerStarted","Data":"abb5fdc349243cbd04cabeb7bf7a1f947bb82348a27808f7ef38cf8e18077d4d"} Dec 04 15:30:30 crc kubenswrapper[4848]: E1204 15:30:30.792512 4848 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.5:54718->38.102.83.5:35999: read tcp 38.102.83.5:54718->38.102.83.5:35999: read: connection reset by peer Dec 04 15:30:42 crc kubenswrapper[4848]: I1204 15:30:42.198858 4848 scope.go:117] "RemoveContainer" containerID="ebf4b9428669e01c1301a64e72614a0485f61b5cd76dd2dca886f34eec536c76" Dec 04 15:30:42 crc kubenswrapper[4848]: I1204 15:30:42.726597 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/crc-debug-lxthz" event={"ID":"8abb634d-2abe-444b-8ab5-04ab450086b7","Type":"ContainerStarted","Data":"6395027a8d8f3bedf2d4dc2b1c1a115f824419d72acdd7f2c1a2a0befb6059aa"} Dec 04 15:30:42 crc kubenswrapper[4848]: I1204 15:30:42.759704 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m2chb/crc-debug-lxthz" podStartSLOduration=1.557397852 podStartE2EDuration="14.759680122s" podCreationTimestamp="2025-12-04 15:30:28 +0000 UTC" firstStartedPulling="2025-12-04 15:30:28.685748401 +0000 UTC m=+6132.628244929" lastFinishedPulling="2025-12-04 15:30:41.888030671 +0000 UTC m=+6145.830527199" observedRunningTime="2025-12-04 15:30:42.741988838 +0000 UTC m=+6146.684485376" watchObservedRunningTime="2025-12-04 15:30:42.759680122 +0000 UTC m=+6146.702176650" Dec 04 15:31:36 crc kubenswrapper[4848]: I1204 15:31:36.360860 4848 generic.go:334] "Generic (PLEG): container finished" podID="8abb634d-2abe-444b-8ab5-04ab450086b7" containerID="6395027a8d8f3bedf2d4dc2b1c1a115f824419d72acdd7f2c1a2a0befb6059aa" exitCode=0 Dec 04 15:31:36 crc kubenswrapper[4848]: I1204 15:31:36.361001 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/crc-debug-lxthz" event={"ID":"8abb634d-2abe-444b-8ab5-04ab450086b7","Type":"ContainerDied","Data":"6395027a8d8f3bedf2d4dc2b1c1a115f824419d72acdd7f2c1a2a0befb6059aa"} Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.551778 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.630766 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m2chb/crc-debug-lxthz"] Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.651972 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m2chb/crc-debug-lxthz"] Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.708185 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8abb634d-2abe-444b-8ab5-04ab450086b7-host\") pod \"8abb634d-2abe-444b-8ab5-04ab450086b7\" (UID: \"8abb634d-2abe-444b-8ab5-04ab450086b7\") " Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.708531 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sc9b\" (UniqueName: \"kubernetes.io/projected/8abb634d-2abe-444b-8ab5-04ab450086b7-kube-api-access-2sc9b\") pod \"8abb634d-2abe-444b-8ab5-04ab450086b7\" (UID: \"8abb634d-2abe-444b-8ab5-04ab450086b7\") " Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.708284 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8abb634d-2abe-444b-8ab5-04ab450086b7-host" (OuterVolumeSpecName: "host") pod "8abb634d-2abe-444b-8ab5-04ab450086b7" (UID: "8abb634d-2abe-444b-8ab5-04ab450086b7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.709528 4848 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8abb634d-2abe-444b-8ab5-04ab450086b7-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.714508 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8abb634d-2abe-444b-8ab5-04ab450086b7-kube-api-access-2sc9b" (OuterVolumeSpecName: "kube-api-access-2sc9b") pod "8abb634d-2abe-444b-8ab5-04ab450086b7" (UID: "8abb634d-2abe-444b-8ab5-04ab450086b7"). InnerVolumeSpecName "kube-api-access-2sc9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:31:37 crc kubenswrapper[4848]: I1204 15:31:37.812173 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sc9b\" (UniqueName: \"kubernetes.io/projected/8abb634d-2abe-444b-8ab5-04ab450086b7-kube-api-access-2sc9b\") on node \"crc\" DevicePath \"\"" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.385656 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abb5fdc349243cbd04cabeb7bf7a1f947bb82348a27808f7ef38cf8e18077d4d" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.385772 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-lxthz" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.408860 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8abb634d-2abe-444b-8ab5-04ab450086b7" path="/var/lib/kubelet/pods/8abb634d-2abe-444b-8ab5-04ab450086b7/volumes" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.783932 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m2chb/crc-debug-lrs8t"] Dec 04 15:31:38 crc kubenswrapper[4848]: E1204 15:31:38.784375 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8abb634d-2abe-444b-8ab5-04ab450086b7" containerName="container-00" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.784391 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="8abb634d-2abe-444b-8ab5-04ab450086b7" containerName="container-00" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.784683 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="8abb634d-2abe-444b-8ab5-04ab450086b7" containerName="container-00" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.785437 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.937632 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-host\") pod \"crc-debug-lrs8t\" (UID: \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\") " pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:38 crc kubenswrapper[4848]: I1204 15:31:38.937680 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtbg8\" (UniqueName: \"kubernetes.io/projected/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-kube-api-access-wtbg8\") pod \"crc-debug-lrs8t\" (UID: \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\") " pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:39 crc kubenswrapper[4848]: I1204 15:31:39.039973 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-host\") pod \"crc-debug-lrs8t\" (UID: \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\") " pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:39 crc kubenswrapper[4848]: I1204 15:31:39.040043 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtbg8\" (UniqueName: \"kubernetes.io/projected/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-kube-api-access-wtbg8\") pod \"crc-debug-lrs8t\" (UID: \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\") " pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:39 crc kubenswrapper[4848]: I1204 15:31:39.040128 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-host\") pod \"crc-debug-lrs8t\" (UID: \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\") " pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:39 crc kubenswrapper[4848]: I1204 15:31:39.063128 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtbg8\" (UniqueName: \"kubernetes.io/projected/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-kube-api-access-wtbg8\") pod \"crc-debug-lrs8t\" (UID: \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\") " pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:39 crc kubenswrapper[4848]: I1204 15:31:39.102628 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:39 crc kubenswrapper[4848]: I1204 15:31:39.396127 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/crc-debug-lrs8t" event={"ID":"6dc9fdbf-b591-43c8-b47a-0813f986ed5f","Type":"ContainerStarted","Data":"597eea1092b86761acb0be40ac605a5a0be0466311c665ab3af6707ae264c317"} Dec 04 15:31:40 crc kubenswrapper[4848]: I1204 15:31:40.426543 4848 generic.go:334] "Generic (PLEG): container finished" podID="6dc9fdbf-b591-43c8-b47a-0813f986ed5f" containerID="3bd5ef924b389175f831c49b9f495ae6ef9a28ddd9930b6a7ef0b74cd47303d5" exitCode=0 Dec 04 15:31:40 crc kubenswrapper[4848]: I1204 15:31:40.426877 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/crc-debug-lrs8t" event={"ID":"6dc9fdbf-b591-43c8-b47a-0813f986ed5f","Type":"ContainerDied","Data":"3bd5ef924b389175f831c49b9f495ae6ef9a28ddd9930b6a7ef0b74cd47303d5"} Dec 04 15:31:41 crc kubenswrapper[4848]: I1204 15:31:41.549311 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:41 crc kubenswrapper[4848]: I1204 15:31:41.711968 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtbg8\" (UniqueName: \"kubernetes.io/projected/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-kube-api-access-wtbg8\") pod \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\" (UID: \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\") " Dec 04 15:31:41 crc kubenswrapper[4848]: I1204 15:31:41.712286 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-host\") pod \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\" (UID: \"6dc9fdbf-b591-43c8-b47a-0813f986ed5f\") " Dec 04 15:31:41 crc kubenswrapper[4848]: I1204 15:31:41.713187 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-host" (OuterVolumeSpecName: "host") pod "6dc9fdbf-b591-43c8-b47a-0813f986ed5f" (UID: "6dc9fdbf-b591-43c8-b47a-0813f986ed5f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:31:41 crc kubenswrapper[4848]: I1204 15:31:41.714178 4848 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:31:41 crc kubenswrapper[4848]: I1204 15:31:41.736945 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-kube-api-access-wtbg8" (OuterVolumeSpecName: "kube-api-access-wtbg8") pod "6dc9fdbf-b591-43c8-b47a-0813f986ed5f" (UID: "6dc9fdbf-b591-43c8-b47a-0813f986ed5f"). InnerVolumeSpecName "kube-api-access-wtbg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:31:41 crc kubenswrapper[4848]: I1204 15:31:41.817907 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtbg8\" (UniqueName: \"kubernetes.io/projected/6dc9fdbf-b591-43c8-b47a-0813f986ed5f-kube-api-access-wtbg8\") on node \"crc\" DevicePath \"\"" Dec 04 15:31:42 crc kubenswrapper[4848]: I1204 15:31:42.448843 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/crc-debug-lrs8t" event={"ID":"6dc9fdbf-b591-43c8-b47a-0813f986ed5f","Type":"ContainerDied","Data":"597eea1092b86761acb0be40ac605a5a0be0466311c665ab3af6707ae264c317"} Dec 04 15:31:42 crc kubenswrapper[4848]: I1204 15:31:42.448886 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="597eea1092b86761acb0be40ac605a5a0be0466311c665ab3af6707ae264c317" Dec 04 15:31:42 crc kubenswrapper[4848]: I1204 15:31:42.448898 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-lrs8t" Dec 04 15:31:43 crc kubenswrapper[4848]: I1204 15:31:43.135864 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m2chb/crc-debug-lrs8t"] Dec 04 15:31:43 crc kubenswrapper[4848]: I1204 15:31:43.148252 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m2chb/crc-debug-lrs8t"] Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.297176 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m2chb/crc-debug-7mqv4"] Dec 04 15:31:44 crc kubenswrapper[4848]: E1204 15:31:44.297878 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dc9fdbf-b591-43c8-b47a-0813f986ed5f" containerName="container-00" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.297891 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dc9fdbf-b591-43c8-b47a-0813f986ed5f" containerName="container-00" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.302342 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dc9fdbf-b591-43c8-b47a-0813f986ed5f" containerName="container-00" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.303160 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.442042 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dc9fdbf-b591-43c8-b47a-0813f986ed5f" path="/var/lib/kubelet/pods/6dc9fdbf-b591-43c8-b47a-0813f986ed5f/volumes" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.475422 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7bf8\" (UniqueName: \"kubernetes.io/projected/9124b5ec-5149-4b4c-8bb2-be9e648d8478-kube-api-access-x7bf8\") pod \"crc-debug-7mqv4\" (UID: \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\") " pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.475499 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9124b5ec-5149-4b4c-8bb2-be9e648d8478-host\") pod \"crc-debug-7mqv4\" (UID: \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\") " pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.579493 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7bf8\" (UniqueName: \"kubernetes.io/projected/9124b5ec-5149-4b4c-8bb2-be9e648d8478-kube-api-access-x7bf8\") pod \"crc-debug-7mqv4\" (UID: \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\") " pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.579625 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9124b5ec-5149-4b4c-8bb2-be9e648d8478-host\") pod \"crc-debug-7mqv4\" (UID: \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\") " pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.579773 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9124b5ec-5149-4b4c-8bb2-be9e648d8478-host\") pod \"crc-debug-7mqv4\" (UID: \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\") " pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.598033 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7bf8\" (UniqueName: \"kubernetes.io/projected/9124b5ec-5149-4b4c-8bb2-be9e648d8478-kube-api-access-x7bf8\") pod \"crc-debug-7mqv4\" (UID: \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\") " pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:44 crc kubenswrapper[4848]: I1204 15:31:44.625805 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:45 crc kubenswrapper[4848]: I1204 15:31:45.537475 4848 generic.go:334] "Generic (PLEG): container finished" podID="9124b5ec-5149-4b4c-8bb2-be9e648d8478" containerID="e3114dbb606c2a01eb788861bca2d98b02d8e4f3de56155717db3623a6dd751b" exitCode=0 Dec 04 15:31:45 crc kubenswrapper[4848]: I1204 15:31:45.537573 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/crc-debug-7mqv4" event={"ID":"9124b5ec-5149-4b4c-8bb2-be9e648d8478","Type":"ContainerDied","Data":"e3114dbb606c2a01eb788861bca2d98b02d8e4f3de56155717db3623a6dd751b"} Dec 04 15:31:45 crc kubenswrapper[4848]: I1204 15:31:45.538135 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/crc-debug-7mqv4" event={"ID":"9124b5ec-5149-4b4c-8bb2-be9e648d8478","Type":"ContainerStarted","Data":"68d371e8646d07db1d8ea121df5d8a06ff09030408124f3c300593967da7c5bc"} Dec 04 15:31:45 crc kubenswrapper[4848]: I1204 15:31:45.584351 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m2chb/crc-debug-7mqv4"] Dec 04 15:31:45 crc kubenswrapper[4848]: I1204 15:31:45.596629 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m2chb/crc-debug-7mqv4"] Dec 04 15:31:46 crc kubenswrapper[4848]: I1204 15:31:46.665966 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:46 crc kubenswrapper[4848]: I1204 15:31:46.726527 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7bf8\" (UniqueName: \"kubernetes.io/projected/9124b5ec-5149-4b4c-8bb2-be9e648d8478-kube-api-access-x7bf8\") pod \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\" (UID: \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\") " Dec 04 15:31:46 crc kubenswrapper[4848]: I1204 15:31:46.726588 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9124b5ec-5149-4b4c-8bb2-be9e648d8478-host\") pod \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\" (UID: \"9124b5ec-5149-4b4c-8bb2-be9e648d8478\") " Dec 04 15:31:46 crc kubenswrapper[4848]: I1204 15:31:46.726772 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9124b5ec-5149-4b4c-8bb2-be9e648d8478-host" (OuterVolumeSpecName: "host") pod "9124b5ec-5149-4b4c-8bb2-be9e648d8478" (UID: "9124b5ec-5149-4b4c-8bb2-be9e648d8478"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:31:46 crc kubenswrapper[4848]: I1204 15:31:46.727129 4848 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9124b5ec-5149-4b4c-8bb2-be9e648d8478-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:31:46 crc kubenswrapper[4848]: I1204 15:31:46.733397 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9124b5ec-5149-4b4c-8bb2-be9e648d8478-kube-api-access-x7bf8" (OuterVolumeSpecName: "kube-api-access-x7bf8") pod "9124b5ec-5149-4b4c-8bb2-be9e648d8478" (UID: "9124b5ec-5149-4b4c-8bb2-be9e648d8478"). InnerVolumeSpecName "kube-api-access-x7bf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:31:46 crc kubenswrapper[4848]: I1204 15:31:46.829131 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7bf8\" (UniqueName: \"kubernetes.io/projected/9124b5ec-5149-4b4c-8bb2-be9e648d8478-kube-api-access-x7bf8\") on node \"crc\" DevicePath \"\"" Dec 04 15:31:47 crc kubenswrapper[4848]: I1204 15:31:47.562706 4848 scope.go:117] "RemoveContainer" containerID="e3114dbb606c2a01eb788861bca2d98b02d8e4f3de56155717db3623a6dd751b" Dec 04 15:31:47 crc kubenswrapper[4848]: I1204 15:31:47.562787 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/crc-debug-7mqv4" Dec 04 15:31:48 crc kubenswrapper[4848]: I1204 15:31:48.406170 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9124b5ec-5149-4b4c-8bb2-be9e648d8478" path="/var/lib/kubelet/pods/9124b5ec-5149-4b4c-8bb2-be9e648d8478/volumes" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.061923 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc646c01-afe4-4b6b-ab39-d8c333cc3ed4/aodh-api/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.178311 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc646c01-afe4-4b6b-ab39-d8c333cc3ed4/aodh-evaluator/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.185767 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc646c01-afe4-4b6b-ab39-d8c333cc3ed4/aodh-listener/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.260829 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc646c01-afe4-4b6b-ab39-d8c333cc3ed4/aodh-notifier/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.386059 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-786894c88b-2c6rg_5e05a6e1-667b-44bd-a532-e91c2245e876/barbican-api/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.389448 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-786894c88b-2c6rg_5e05a6e1-667b-44bd-a532-e91c2245e876/barbican-api-log/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.533089 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58957fbb8b-n55cg_c232050c-845f-4a98-b79d-1788830cfd67/barbican-keystone-listener/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.676744 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58957fbb8b-n55cg_c232050c-845f-4a98-b79d-1788830cfd67/barbican-keystone-listener-log/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.711407 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b699c955-dtjdb_051c8366-c751-423e-a7c8-4d69b6a5897d/barbican-worker/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.774871 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b699c955-dtjdb_051c8366-c751-423e-a7c8-4d69b6a5897d/barbican-worker-log/0.log" Dec 04 15:32:11 crc kubenswrapper[4848]: I1204 15:32:11.908419 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5_17b7cb97-4979-468f-89be-d14529b93be8/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.105026 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40c4de59-893b-42be-b67d-f3f067b410ac/ceilometer-notification-agent/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.114083 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40c4de59-893b-42be-b67d-f3f067b410ac/ceilometer-central-agent/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.152536 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40c4de59-893b-42be-b67d-f3f067b410ac/proxy-httpd/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.203773 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40c4de59-893b-42be-b67d-f3f067b410ac/sg-core/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.378677 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e0893eec-f2cb-41b8-a301-005dbf326afe/cinder-api-log/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.442766 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e0893eec-f2cb-41b8-a301-005dbf326afe/cinder-api/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.616293 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c93bea85-4d93-4003-b146-70a07e1edd92/cinder-scheduler/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.697214 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9_fbf57421-8b0a-42bd-a682-65afd34f39bb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.699030 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c93bea85-4d93-4003-b146-70a07e1edd92/probe/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.956544 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt_8c603058-8c35-4c40-b700-2c358d0a5df7/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:12 crc kubenswrapper[4848]: I1204 15:32:12.980681 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-g7qqg_7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd/init/0.log" Dec 04 15:32:13 crc kubenswrapper[4848]: I1204 15:32:13.081472 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-g7qqg_7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd/init/0.log" Dec 04 15:32:13 crc kubenswrapper[4848]: I1204 15:32:13.195484 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-g7qqg_7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd/dnsmasq-dns/0.log" Dec 04 15:32:13 crc kubenswrapper[4848]: I1204 15:32:13.216480 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp_9d69a925-9c1f-4405-9300-e6c4c3cada23/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:13 crc kubenswrapper[4848]: I1204 15:32:13.414755 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d5008baa-3c8b-4048-a87e-4e6f3e5d2327/glance-log/0.log" Dec 04 15:32:13 crc kubenswrapper[4848]: I1204 15:32:13.432644 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d5008baa-3c8b-4048-a87e-4e6f3e5d2327/glance-httpd/0.log" Dec 04 15:32:13 crc kubenswrapper[4848]: I1204 15:32:13.596785 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0c33bbab-5535-4d34-8b87-90c7e1549281/glance-httpd/0.log" Dec 04 15:32:13 crc kubenswrapper[4848]: I1204 15:32:13.628421 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0c33bbab-5535-4d34-8b87-90c7e1549281/glance-log/0.log" Dec 04 15:32:14 crc kubenswrapper[4848]: I1204 15:32:14.158986 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-55957f9948-8gknv_01ecc04b-d6b6-4182-ba10-0b2d14b6bef5/heat-engine/0.log" Dec 04 15:32:14 crc kubenswrapper[4848]: I1204 15:32:14.420720 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp_fe59edb1-8c86-40a3-b68a-7f597af859de/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:14 crc kubenswrapper[4848]: I1204 15:32:14.543336 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6c746cbd45-fc8bq_4ab14888-fa03-4bbd-a567-95e9078fbe3b/heat-api/0.log" Dec 04 15:32:14 crc kubenswrapper[4848]: I1204 15:32:14.555966 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-6b8f46df8d-vvlcd_cfffd13e-a23e-41bd-b830-0feab48494ec/heat-cfnapi/0.log" Dec 04 15:32:14 crc kubenswrapper[4848]: I1204 15:32:14.648070 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-gbtwr_51fab24f-2ea8-430c-810b-a538832876de/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:14 crc kubenswrapper[4848]: I1204 15:32:14.850130 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29414341-72l4q_9fe270e6-d360-4e3f-851e-3ab6b76b92a9/keystone-cron/0.log" Dec 04 15:32:15 crc kubenswrapper[4848]: I1204 15:32:15.003893 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1d88c8b4-d31b-45d7-8ae5-99b50d716ae0/kube-state-metrics/0.log" Dec 04 15:32:15 crc kubenswrapper[4848]: I1204 15:32:15.162244 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-6lph2_3146cd9a-c65d-4942-bb4d-ee0951403d87/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:15 crc kubenswrapper[4848]: I1204 15:32:15.178572 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5b74ff8c74-nm59b_6a28b540-21e3-43eb-9b64-215d661d4721/keystone-api/0.log" Dec 04 15:32:15 crc kubenswrapper[4848]: I1204 15:32:15.208828 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-9zcjg_d0800b88-1124-4b08-a53c-ba88de0acc98/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:15 crc kubenswrapper[4848]: I1204 15:32:15.396136 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_2ddeb253-877f-4e3e-8d1c-b51c1621bd15/mysqld-exporter/0.log" Dec 04 15:32:15 crc kubenswrapper[4848]: I1204 15:32:15.867700 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr_b1231a87-3a32-4b52-a5b0-d546f15523e1/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:15 crc kubenswrapper[4848]: I1204 15:32:15.962295 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b6b4654c9-hg8j8_59773efa-83b9-4645-8187-fc0ba4c1d70c/neutron-api/0.log" Dec 04 15:32:16 crc kubenswrapper[4848]: I1204 15:32:16.016216 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b6b4654c9-hg8j8_59773efa-83b9-4645-8187-fc0ba4c1d70c/neutron-httpd/0.log" Dec 04 15:32:16 crc kubenswrapper[4848]: I1204 15:32:16.457519 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7ae44733-c46b-4ed7-9371-377d026bc216/nova-cell0-conductor-conductor/0.log" Dec 04 15:32:16 crc kubenswrapper[4848]: I1204 15:32:16.817329 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fd31580e-d25b-49af-a7c6-2eb44a77af46/nova-cell1-conductor-conductor/0.log" Dec 04 15:32:16 crc kubenswrapper[4848]: I1204 15:32:16.833824 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d9781260-81d8-4950-8388-00ae237a6f7f/nova-api-log/0.log" Dec 04 15:32:17 crc kubenswrapper[4848]: I1204 15:32:17.104087 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_1c49601a-f35b-41cb-bbb6-d7a4dd0542c0/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 15:32:17 crc kubenswrapper[4848]: I1204 15:32:17.146989 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-v9fxp_86fbd978-880e-415b-ad1e-e367a0c0079b/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:17 crc kubenswrapper[4848]: I1204 15:32:17.225893 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d9781260-81d8-4950-8388-00ae237a6f7f/nova-api-api/0.log" Dec 04 15:32:17 crc kubenswrapper[4848]: I1204 15:32:17.377975 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_42fe17d3-9143-40e0-ab66-074885ccd028/nova-metadata-log/0.log" Dec 04 15:32:17 crc kubenswrapper[4848]: I1204 15:32:17.727327 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7fa9a476-cd4b-4258-ba07-0ef2888fa4e3/nova-scheduler-scheduler/0.log" Dec 04 15:32:17 crc kubenswrapper[4848]: I1204 15:32:17.783931 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2/mysql-bootstrap/0.log" Dec 04 15:32:17 crc kubenswrapper[4848]: I1204 15:32:17.974409 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2/mysql-bootstrap/0.log" Dec 04 15:32:17 crc kubenswrapper[4848]: I1204 15:32:17.990491 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2/galera/0.log" Dec 04 15:32:18 crc kubenswrapper[4848]: I1204 15:32:18.228122 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52/mysql-bootstrap/0.log" Dec 04 15:32:18 crc kubenswrapper[4848]: I1204 15:32:18.405312 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52/mysql-bootstrap/0.log" Dec 04 15:32:18 crc kubenswrapper[4848]: I1204 15:32:18.443354 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52/galera/0.log" Dec 04 15:32:18 crc kubenswrapper[4848]: I1204 15:32:18.591029 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ec3f1251-eb61-4b0e-b534-177c7d928404/openstackclient/0.log" Dec 04 15:32:18 crc kubenswrapper[4848]: I1204 15:32:18.695453 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-b9f5p_caf1c526-2dfe-4c81-9846-c122113ebaf1/openstack-network-exporter/0.log" Dec 04 15:32:18 crc kubenswrapper[4848]: I1204 15:32:18.875716 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cpcmr_0ef872ad-6863-45f9-ac22-86841dea2bbf/ovsdb-server-init/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.073577 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cpcmr_0ef872ad-6863-45f9-ac22-86841dea2bbf/ovsdb-server-init/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.083463 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cpcmr_0ef872ad-6863-45f9-ac22-86841dea2bbf/ovs-vswitchd/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.085365 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cpcmr_0ef872ad-6863-45f9-ac22-86841dea2bbf/ovsdb-server/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.334617 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-rg7xl_ef9a0101-72b3-4879-a93b-f58362d175ed/ovn-controller/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.607913 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c7da8d50-9ec2-4b1d-b373-e65541b02306/openstack-network-exporter/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.611624 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-l9mmb_54ce5df0-f586-463e-ad53-6ddc68e8ce90/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.623443 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_42fe17d3-9143-40e0-ab66-074885ccd028/nova-metadata-metadata/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.841209 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_12b7748e-3169-49a0-a634-8a485f7690f4/openstack-network-exporter/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.854097 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c7da8d50-9ec2-4b1d-b373-e65541b02306/ovn-northd/0.log" Dec 04 15:32:19 crc kubenswrapper[4848]: I1204 15:32:19.863177 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_12b7748e-3169-49a0-a634-8a485f7690f4/ovsdbserver-nb/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.023283 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_97bf1fd2-3a36-418d-826a-6dc60889b70c/openstack-network-exporter/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.094903 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_97bf1fd2-3a36-418d-826a-6dc60889b70c/ovsdbserver-sb/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.316037 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/init-config-reloader/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.396673 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8766cc68-d2824_7c8a40cb-aac2-4a08-96f3-79c97ad82007/placement-api/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.455326 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8766cc68-d2824_7c8a40cb-aac2-4a08-96f3-79c97ad82007/placement-log/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.589588 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/init-config-reloader/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.631943 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/config-reloader/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.644431 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/prometheus/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.679781 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/thanos-sidecar/0.log" Dec 04 15:32:20 crc kubenswrapper[4848]: I1204 15:32:20.841005 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6c94605c-497c-47b2-b606-22f72e1bd0cc/setup-container/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.084768 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6c94605c-497c-47b2-b606-22f72e1bd0cc/setup-container/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.121349 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a33214fe-5596-499a-9563-0931c49821c3/setup-container/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.125392 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6c94605c-497c-47b2-b606-22f72e1bd0cc/rabbitmq/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.362909 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a33214fe-5596-499a-9563-0931c49821c3/setup-container/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.394151 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a33214fe-5596-499a-9563-0931c49821c3/rabbitmq/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.426961 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf_2af1d16b-4257-4fcd-b6fa-dc37b1313e40/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.551454 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-kwkbw_25e07c34-72be-4c3a-8432-3fb295a46d57/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.677053 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz_66ec76cb-b6b1-4846-83d3-58cf60fd2d3d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.876244 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dfmnc_5b8a5833-af2e-4fef-9bfa-bbcec7064538/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:21 crc kubenswrapper[4848]: I1204 15:32:21.926617 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dxn92_7ae92c99-2f33-44d5-a478-9a685afd0f89/ssh-known-hosts-edpm-deployment/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.145115 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b65c5f4cf-q22f8_acf82280-f41a-460a-84ab-639c541670fc/proxy-server/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.320618 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gplgt_ff826431-be61-4dea-82e7-b6849d838e52/swift-ring-rebalance/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.380569 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b65c5f4cf-q22f8_acf82280-f41a-460a-84ab-639c541670fc/proxy-httpd/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.446490 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/account-auditor/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.531686 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/account-reaper/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.865738 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/account-server/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.888895 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/account-replicator/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.945045 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/container-auditor/0.log" Dec 04 15:32:22 crc kubenswrapper[4848]: I1204 15:32:22.953242 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/container-replicator/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.060320 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/container-server/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.145648 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/container-updater/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.181935 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-auditor/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.206616 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-expirer/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.335708 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-replicator/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.402975 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-server/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.404864 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-updater/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.417239 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/rsync/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.527869 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/swift-recon-cron/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.720688 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs_d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:23 crc kubenswrapper[4848]: I1204 15:32:23.813340 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg_0e112a8b-6965-4e32-bb39-4ff08519bdee/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:24 crc kubenswrapper[4848]: I1204 15:32:24.132184 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_3d34d919-565d-49be-8170-1bec9816c2c9/test-operator-logs-container/0.log" Dec 04 15:32:24 crc kubenswrapper[4848]: I1204 15:32:24.240597 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt_bae5190f-858d-4fdc-8116-78911f113311/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:32:25 crc kubenswrapper[4848]: I1204 15:32:25.154684 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a/tempest-tests-tempest-tests-runner/0.log" Dec 04 15:32:39 crc kubenswrapper[4848]: I1204 15:32:39.663929 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_add25985-bc0c-48ee-9f42-4404503638b5/memcached/0.log" Dec 04 15:32:44 crc kubenswrapper[4848]: I1204 15:32:44.314304 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:32:44 crc kubenswrapper[4848]: I1204 15:32:44.314905 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:32:52 crc kubenswrapper[4848]: I1204 15:32:52.550570 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-65bvj_4ad79d92-0097-4c46-930a-75489fad9915/kube-rbac-proxy/0.log" Dec 04 15:32:52 crc kubenswrapper[4848]: I1204 15:32:52.646507 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-65bvj_4ad79d92-0097-4c46-930a-75489fad9915/manager/0.log" Dec 04 15:32:52 crc kubenswrapper[4848]: I1204 15:32:52.825014 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-9kj8j_038e3294-a0db-4c23-8ef2-ed6a7e617a3f/kube-rbac-proxy/0.log" Dec 04 15:32:52 crc kubenswrapper[4848]: I1204 15:32:52.932219 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-9kj8j_038e3294-a0db-4c23-8ef2-ed6a7e617a3f/manager/0.log" Dec 04 15:32:52 crc kubenswrapper[4848]: I1204 15:32:52.998372 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-pv6z2_7a7135b8-89ed-4a8a-a09e-7e343bcd30ba/kube-rbac-proxy/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.061835 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-pv6z2_7a7135b8-89ed-4a8a-a09e-7e343bcd30ba/manager/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.172908 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/util/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.264978 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/util/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.319654 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/pull/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.338488 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/pull/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.561899 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/util/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.567803 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/extract/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.581754 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/pull/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.748835 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-zljs9_7b00aa45-21c2-4be2-b702-8f191fc12b24/kube-rbac-proxy/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.815779 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-zljs9_7b00aa45-21c2-4be2-b702-8f191fc12b24/manager/0.log" Dec 04 15:32:53 crc kubenswrapper[4848]: I1204 15:32:53.823706 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4zcdk_1db7d065-38b0-4ef5-a468-660c4073f7f6/kube-rbac-proxy/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.005936 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4zcdk_1db7d065-38b0-4ef5-a468-660c4073f7f6/manager/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.028865 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pkb5l_a9c2feb0-4b79-4adc-b7a6-885a5a33c244/kube-rbac-proxy/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.049601 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pkb5l_a9c2feb0-4b79-4adc-b7a6-885a5a33c244/manager/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.210326 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-7g8w2_c22be20e-3619-45f0-944a-eff494ac7f1a/kube-rbac-proxy/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.374103 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-7g8w2_c22be20e-3619-45f0-944a-eff494ac7f1a/manager/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.408122 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-4hmhf_2fc8dada-7c7b-494c-ae6d-72c2737c6f66/kube-rbac-proxy/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.434539 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-4hmhf_2fc8dada-7c7b-494c-ae6d-72c2737c6f66/manager/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.596088 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kvfn2_de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a/kube-rbac-proxy/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.688416 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kvfn2_de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a/manager/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.717772 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-925th_2c849714-e889-446a-9b2c-6862cdbaf422/kube-rbac-proxy/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.771535 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-925th_2c849714-e889-446a-9b2c-6862cdbaf422/manager/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.875447 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-9sqtl_82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93/kube-rbac-proxy/0.log" Dec 04 15:32:54 crc kubenswrapper[4848]: I1204 15:32:54.933286 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-9sqtl_82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93/manager/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.061268 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rr887_16e8aa39-31cb-4051-bdcf-797b4d5792c3/kube-rbac-proxy/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.115709 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rr887_16e8aa39-31cb-4051-bdcf-797b4d5792c3/manager/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.235812 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-dvhkz_9e10ea48-b397-4f33-9da2-1a82a2987a9e/kube-rbac-proxy/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.342504 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-dvhkz_9e10ea48-b397-4f33-9da2-1a82a2987a9e/manager/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.363230 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-gdsqn_fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7/kube-rbac-proxy/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.454116 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-gdsqn_fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7/manager/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.537200 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd46t68x_a60c0a88-4168-4971-8bdd-638030697efb/kube-rbac-proxy/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.585709 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd46t68x_a60c0a88-4168-4971-8bdd-638030697efb/manager/0.log" Dec 04 15:32:55 crc kubenswrapper[4848]: I1204 15:32:55.939271 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-799774d886-rj5dt_6641bd6c-020c-469a-8439-4953ffa85ee9/operator/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.019439 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-st8tz_409845c7-53dc-40ea-902b-b36b35b7d686/registry-server/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.129026 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-8hd8b_14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6/kube-rbac-proxy/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.365877 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-8hd8b_14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6/manager/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.454178 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c4j2k_f15b3026-2d34-491c-a37f-c0053e3171ee/kube-rbac-proxy/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.605229 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5l2qh_0488ea60-3643-4e40-94cd-555ac6c93043/operator/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.607316 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c4j2k_f15b3026-2d34-491c-a37f-c0053e3171ee/manager/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.853475 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-kdftj_82a80725-8983-447c-8f98-d4b298ae8e7e/kube-rbac-proxy/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.877492 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-kdftj_82a80725-8983-447c-8f98-d4b298ae8e7e/manager/0.log" Dec 04 15:32:56 crc kubenswrapper[4848]: I1204 15:32:56.977335 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f78686648-gktm6_2c5d62fd-6700-4804-ba6d-4d9884b9ea02/kube-rbac-proxy/0.log" Dec 04 15:32:57 crc kubenswrapper[4848]: I1204 15:32:57.148434 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z2wqt_12439af0-d3a3-47d8-9626-b7e357af2100/kube-rbac-proxy/0.log" Dec 04 15:32:57 crc kubenswrapper[4848]: I1204 15:32:57.150038 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-576c48d787-gx8wm_12302e73-00be-4ffb-9cda-3ac6126e2f0c/manager/0.log" Dec 04 15:32:57 crc kubenswrapper[4848]: I1204 15:32:57.220669 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z2wqt_12439af0-d3a3-47d8-9626-b7e357af2100/manager/0.log" Dec 04 15:32:57 crc kubenswrapper[4848]: I1204 15:32:57.388865 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f78686648-gktm6_2c5d62fd-6700-4804-ba6d-4d9884b9ea02/manager/0.log" Dec 04 15:32:57 crc kubenswrapper[4848]: I1204 15:32:57.412019 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-sr98b_5382b606-8cea-4b64-98ac-950f99c2a03d/kube-rbac-proxy/0.log" Dec 04 15:32:57 crc kubenswrapper[4848]: I1204 15:32:57.426004 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-sr98b_5382b606-8cea-4b64-98ac-950f99c2a03d/manager/0.log" Dec 04 15:33:14 crc kubenswrapper[4848]: I1204 15:33:14.314282 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:33:14 crc kubenswrapper[4848]: I1204 15:33:14.314791 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:33:16 crc kubenswrapper[4848]: I1204 15:33:16.713919 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qx7kz_41654867-fa7f-43c7-b672-c3eb6500b16c/control-plane-machine-set-operator/0.log" Dec 04 15:33:16 crc kubenswrapper[4848]: I1204 15:33:16.918311 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tk6sb_e8f7ac81-45f7-4446-a6df-78123ca7226b/machine-api-operator/0.log" Dec 04 15:33:16 crc kubenswrapper[4848]: I1204 15:33:16.928668 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tk6sb_e8f7ac81-45f7-4446-a6df-78123ca7226b/kube-rbac-proxy/0.log" Dec 04 15:33:29 crc kubenswrapper[4848]: I1204 15:33:29.428668 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-wcntc_327abad7-b085-4c01-8ecd-e287bda9f6ee/cert-manager-controller/0.log" Dec 04 15:33:29 crc kubenswrapper[4848]: I1204 15:33:29.571093 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-r78j7_d708a779-4728-4230-ba68-02eb26b4dc72/cert-manager-cainjector/0.log" Dec 04 15:33:29 crc kubenswrapper[4848]: I1204 15:33:29.627380 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-76vk7_7e550e0e-d6e6-4786-ae2b-46a4f1534877/cert-manager-webhook/0.log" Dec 04 15:33:42 crc kubenswrapper[4848]: I1204 15:33:42.806974 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-qj7vf_357ea823-946e-407d-8940-b430c72e324e/nmstate-console-plugin/0.log" Dec 04 15:33:43 crc kubenswrapper[4848]: I1204 15:33:43.031655 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-rggz7_973c2430-1dce-45e2-b08f-f87439b80cce/nmstate-handler/0.log" Dec 04 15:33:43 crc kubenswrapper[4848]: I1204 15:33:43.079852 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-4nd5h_afaef749-1755-4749-bb8f-64b3fcb3f768/kube-rbac-proxy/0.log" Dec 04 15:33:43 crc kubenswrapper[4848]: I1204 15:33:43.131790 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-4nd5h_afaef749-1755-4749-bb8f-64b3fcb3f768/nmstate-metrics/0.log" Dec 04 15:33:43 crc kubenswrapper[4848]: I1204 15:33:43.320679 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-69mx9_33caa45e-2061-41cc-8dca-94a382ffa50a/nmstate-operator/0.log" Dec 04 15:33:43 crc kubenswrapper[4848]: I1204 15:33:43.383919 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-sjwbx_ba09ba52-535f-41bc-adb2-210ca566a77a/nmstate-webhook/0.log" Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.314628 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.314971 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.315022 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.315972 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e8b1d4f65c5c25aa19d0b9bebebf377fde86d7d15e1e29fa678a12b8a93b1e9"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.316315 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://6e8b1d4f65c5c25aa19d0b9bebebf377fde86d7d15e1e29fa678a12b8a93b1e9" gracePeriod=600 Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.976524 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="6e8b1d4f65c5c25aa19d0b9bebebf377fde86d7d15e1e29fa678a12b8a93b1e9" exitCode=0 Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.976602 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"6e8b1d4f65c5c25aa19d0b9bebebf377fde86d7d15e1e29fa678a12b8a93b1e9"} Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.976998 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760"} Dec 04 15:33:44 crc kubenswrapper[4848]: I1204 15:33:44.977020 4848 scope.go:117] "RemoveContainer" containerID="2eae851b686dd8095ecce3f4d409fc579b805dfbfb6c257a2f94de40b0bab844" Dec 04 15:33:57 crc kubenswrapper[4848]: I1204 15:33:57.255147 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7ddb79586-x6vmg_461d1ece-26ff-40ed-a92c-f21300adeec7/manager/0.log" Dec 04 15:33:57 crc kubenswrapper[4848]: I1204 15:33:57.284703 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7ddb79586-x6vmg_461d1ece-26ff-40ed-a92c-f21300adeec7/kube-rbac-proxy/0.log" Dec 04 15:34:11 crc kubenswrapper[4848]: I1204 15:34:11.852403 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-lgmcj_5d3f5a36-9e68-4f3d-94b8-8cb59192211b/cluster-logging-operator/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.066123 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-wgpm5_13d80a8f-6a05-4fdb-99bd-6aaf5f563f87/collector/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.131117 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_b53d48d7-e9d7-442d-b2b8-044a7eb9e692/loki-compactor/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.297845 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-mkvbk_cf9c3660-b2e2-4fac-a328-832144acf92c/loki-distributor/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.363971 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-7db4f94957-p9qcz_3457940c-ae43-4e49-b616-a450d524ffd9/gateway/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.457322 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-7db4f94957-p9qcz_3457940c-ae43-4e49-b616-a450d524ffd9/opa/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.591670 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-7db4f94957-qqjgk_bf4fb005-0de1-45b7-95a1-0fc7783e3a72/gateway/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.632272 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-7db4f94957-qqjgk_bf4fb005-0de1-45b7-95a1-0fc7783e3a72/opa/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.762937 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_25f82b9f-ffd8-47b2-9abe-593234820cb5/loki-index-gateway/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.908455 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_19d88d7c-8641-499e-8fc1-8864af7feb3c/loki-ingester/0.log" Dec 04 15:34:12 crc kubenswrapper[4848]: I1204 15:34:12.956073 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-g78k5_6d38276f-2908-496a-aa46-175984204c26/loki-querier/0.log" Dec 04 15:34:13 crc kubenswrapper[4848]: I1204 15:34:13.131051 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-htkln_d659950d-e89c-45fa-90c5-4526593ed819/loki-query-frontend/0.log" Dec 04 15:34:27 crc kubenswrapper[4848]: I1204 15:34:27.398459 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-h8bgt_e8c600d3-9a0f-4cab-9581-e75fa003c19d/kube-rbac-proxy/0.log" Dec 04 15:34:27 crc kubenswrapper[4848]: I1204 15:34:27.562732 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-h8bgt_e8c600d3-9a0f-4cab-9581-e75fa003c19d/controller/0.log" Dec 04 15:34:27 crc kubenswrapper[4848]: I1204 15:34:27.655730 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-frr-files/0.log" Dec 04 15:34:27 crc kubenswrapper[4848]: I1204 15:34:27.828869 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-frr-files/0.log" Dec 04 15:34:27 crc kubenswrapper[4848]: I1204 15:34:27.846185 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-reloader/0.log" Dec 04 15:34:27 crc kubenswrapper[4848]: I1204 15:34:27.874463 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-metrics/0.log" Dec 04 15:34:27 crc kubenswrapper[4848]: I1204 15:34:27.895892 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-reloader/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.090981 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-reloader/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.130053 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-frr-files/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.134691 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-metrics/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.174240 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-metrics/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.387888 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-frr-files/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.403206 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-metrics/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.410554 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-reloader/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.422314 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/controller/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.618261 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/kube-rbac-proxy-frr/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.620816 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/kube-rbac-proxy/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.641325 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/frr-metrics/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.835285 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/reloader/0.log" Dec 04 15:34:28 crc kubenswrapper[4848]: I1204 15:34:28.856153 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-gpdjr_054c1d24-8144-490d-b917-391860e5704d/frr-k8s-webhook-server/0.log" Dec 04 15:34:29 crc kubenswrapper[4848]: I1204 15:34:29.099410 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-657d4c7649-z2tns_fe476aa8-681e-4dfb-9e65-58910c24c9cc/manager/0.log" Dec 04 15:34:29 crc kubenswrapper[4848]: I1204 15:34:29.316342 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c56865d68-9mdfg_ab8ac511-522e-4116-9c7b-ab61a4a6e018/webhook-server/0.log" Dec 04 15:34:29 crc kubenswrapper[4848]: I1204 15:34:29.445329 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pxcmt_e1998e0d-8a5c-4a7a-8008-598967ed2da0/kube-rbac-proxy/0.log" Dec 04 15:34:30 crc kubenswrapper[4848]: I1204 15:34:30.124152 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pxcmt_e1998e0d-8a5c-4a7a-8008-598967ed2da0/speaker/0.log" Dec 04 15:34:30 crc kubenswrapper[4848]: I1204 15:34:30.399243 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/frr/0.log" Dec 04 15:34:42 crc kubenswrapper[4848]: I1204 15:34:42.879195 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/util/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.025813 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/util/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.027233 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/pull/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.056932 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/pull/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.279831 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/util/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.304835 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/extract/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.308540 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/pull/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.456542 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/util/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.628749 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/util/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.660517 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/pull/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.660687 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/pull/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.833852 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/pull/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.866893 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/util/0.log" Dec 04 15:34:43 crc kubenswrapper[4848]: I1204 15:34:43.904831 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/extract/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.041697 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/util/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.217870 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/util/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.226497 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/pull/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.231584 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/pull/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.386438 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/pull/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.418659 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/util/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.456177 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/extract/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.590997 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/util/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.793124 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/pull/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.812015 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/util/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.845468 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/pull/0.log" Dec 04 15:34:44 crc kubenswrapper[4848]: I1204 15:34:44.980530 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/util/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.025345 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/pull/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.053574 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/extract/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.177044 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/util/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.417218 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/pull/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.427526 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/pull/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.496990 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/util/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.799189 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/util/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.799905 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/pull/0.log" Dec 04 15:34:45 crc kubenswrapper[4848]: I1204 15:34:45.864664 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/extract/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.040486 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-utilities/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.277632 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-utilities/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.299399 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-content/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.358883 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-content/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.522849 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-utilities/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.576936 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-content/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.715564 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-utilities/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.904891 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/registry-server/0.log" Dec 04 15:34:46 crc kubenswrapper[4848]: I1204 15:34:46.982622 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-utilities/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.014741 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-content/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.028967 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-content/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.160878 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-utilities/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.197394 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-content/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.306564 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-nq9tw_cf33e30f-cf18-4cdd-8b35-84f835998ffd/marketplace-operator/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.432333 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-utilities/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.644708 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-content/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.665214 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-utilities/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.732573 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-content/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.924447 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-content/0.log" Dec 04 15:34:47 crc kubenswrapper[4848]: I1204 15:34:47.936842 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-utilities/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.192921 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-utilities/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.292636 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/registry-server/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.385695 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/registry-server/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.395974 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-utilities/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.412065 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-content/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.444501 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-content/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.627795 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-utilities/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.630114 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-content/0.log" Dec 04 15:34:48 crc kubenswrapper[4848]: I1204 15:34:48.871338 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/registry-server/0.log" Dec 04 15:35:01 crc kubenswrapper[4848]: I1204 15:35:01.440040 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-mwx95_41de33f3-4fb3-48fa-988b-cb6f059cab9e/prometheus-operator/0.log" Dec 04 15:35:01 crc kubenswrapper[4848]: I1204 15:35:01.606707 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_be2f3162-5720-4171-82c9-4df4a4244ec8/prometheus-operator-admission-webhook/0.log" Dec 04 15:35:01 crc kubenswrapper[4848]: I1204 15:35:01.676157 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_202cc8e2-1429-410d-8c7f-36062221f810/prometheus-operator-admission-webhook/0.log" Dec 04 15:35:01 crc kubenswrapper[4848]: I1204 15:35:01.818933 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-fl7j8_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb/operator/0.log" Dec 04 15:35:01 crc kubenswrapper[4848]: I1204 15:35:01.878970 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-49kl7_8efe0158-56d4-4526-9683-263fb1ce6f23/observability-ui-dashboards/0.log" Dec 04 15:35:02 crc kubenswrapper[4848]: I1204 15:35:02.021802 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-tn52h_d6694eaf-b146-407c-95b2-c52ccc355a77/perses-operator/0.log" Dec 04 15:35:14 crc kubenswrapper[4848]: I1204 15:35:14.274500 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7ddb79586-x6vmg_461d1ece-26ff-40ed-a92c-f21300adeec7/kube-rbac-proxy/0.log" Dec 04 15:35:14 crc kubenswrapper[4848]: I1204 15:35:14.320575 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7ddb79586-x6vmg_461d1ece-26ff-40ed-a92c-f21300adeec7/manager/0.log" Dec 04 15:35:25 crc kubenswrapper[4848]: E1204 15:35:25.456315 4848 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.5:48938->38.102.83.5:35999: read tcp 38.102.83.5:48938->38.102.83.5:35999: read: connection reset by peer Dec 04 15:35:29 crc kubenswrapper[4848]: E1204 15:35:29.529108 4848 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.5:51460->38.102.83.5:35999: write tcp 38.102.83.5:51460->38.102.83.5:35999: write: broken pipe Dec 04 15:35:44 crc kubenswrapper[4848]: I1204 15:35:44.316941 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:35:44 crc kubenswrapper[4848]: I1204 15:35:44.317847 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:35:56 crc kubenswrapper[4848]: I1204 15:35:56.937893 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-974ms"] Dec 04 15:35:56 crc kubenswrapper[4848]: E1204 15:35:56.939122 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9124b5ec-5149-4b4c-8bb2-be9e648d8478" containerName="container-00" Dec 04 15:35:56 crc kubenswrapper[4848]: I1204 15:35:56.939140 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="9124b5ec-5149-4b4c-8bb2-be9e648d8478" containerName="container-00" Dec 04 15:35:56 crc kubenswrapper[4848]: I1204 15:35:56.939497 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="9124b5ec-5149-4b4c-8bb2-be9e648d8478" containerName="container-00" Dec 04 15:35:56 crc kubenswrapper[4848]: I1204 15:35:56.943404 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:56 crc kubenswrapper[4848]: I1204 15:35:56.972631 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-974ms"] Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.022304 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-utilities\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.022420 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcgr2\" (UniqueName: \"kubernetes.io/projected/78a42305-57db-4198-81b0-4f419c6d4544-kube-api-access-mcgr2\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.022529 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-catalog-content\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.125311 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcgr2\" (UniqueName: \"kubernetes.io/projected/78a42305-57db-4198-81b0-4f419c6d4544-kube-api-access-mcgr2\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.125440 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-catalog-content\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.125529 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-utilities\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.125969 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-catalog-content\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.125988 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-utilities\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.153813 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcgr2\" (UniqueName: \"kubernetes.io/projected/78a42305-57db-4198-81b0-4f419c6d4544-kube-api-access-mcgr2\") pod \"certified-operators-974ms\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.273169 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:35:57 crc kubenswrapper[4848]: I1204 15:35:57.939426 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-974ms"] Dec 04 15:35:57 crc kubenswrapper[4848]: W1204 15:35:57.941141 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78a42305_57db_4198_81b0_4f419c6d4544.slice/crio-ec44ed255d1443438fe089361798c560598b1f91cc55883bb603be629652993d WatchSource:0}: Error finding container ec44ed255d1443438fe089361798c560598b1f91cc55883bb603be629652993d: Status 404 returned error can't find the container with id ec44ed255d1443438fe089361798c560598b1f91cc55883bb603be629652993d Dec 04 15:35:58 crc kubenswrapper[4848]: I1204 15:35:58.523522 4848 generic.go:334] "Generic (PLEG): container finished" podID="78a42305-57db-4198-81b0-4f419c6d4544" containerID="9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d" exitCode=0 Dec 04 15:35:58 crc kubenswrapper[4848]: I1204 15:35:58.523605 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-974ms" event={"ID":"78a42305-57db-4198-81b0-4f419c6d4544","Type":"ContainerDied","Data":"9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d"} Dec 04 15:35:58 crc kubenswrapper[4848]: I1204 15:35:58.523829 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-974ms" event={"ID":"78a42305-57db-4198-81b0-4f419c6d4544","Type":"ContainerStarted","Data":"ec44ed255d1443438fe089361798c560598b1f91cc55883bb603be629652993d"} Dec 04 15:35:58 crc kubenswrapper[4848]: I1204 15:35:58.528413 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:36:00 crc kubenswrapper[4848]: I1204 15:36:00.545656 4848 generic.go:334] "Generic (PLEG): container finished" podID="78a42305-57db-4198-81b0-4f419c6d4544" containerID="0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c" exitCode=0 Dec 04 15:36:00 crc kubenswrapper[4848]: I1204 15:36:00.545743 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-974ms" event={"ID":"78a42305-57db-4198-81b0-4f419c6d4544","Type":"ContainerDied","Data":"0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c"} Dec 04 15:36:01 crc kubenswrapper[4848]: I1204 15:36:01.559283 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-974ms" event={"ID":"78a42305-57db-4198-81b0-4f419c6d4544","Type":"ContainerStarted","Data":"eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f"} Dec 04 15:36:01 crc kubenswrapper[4848]: I1204 15:36:01.588193 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-974ms" podStartSLOduration=3.00029042 podStartE2EDuration="5.58817012s" podCreationTimestamp="2025-12-04 15:35:56 +0000 UTC" firstStartedPulling="2025-12-04 15:35:58.527025142 +0000 UTC m=+6462.469521680" lastFinishedPulling="2025-12-04 15:36:01.114904842 +0000 UTC m=+6465.057401380" observedRunningTime="2025-12-04 15:36:01.581508357 +0000 UTC m=+6465.524004895" watchObservedRunningTime="2025-12-04 15:36:01.58817012 +0000 UTC m=+6465.530666648" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.695221 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xfpx5"] Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.699616 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.731076 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfpx5"] Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.841384 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckw6t\" (UniqueName: \"kubernetes.io/projected/a60347a1-afa5-4191-9daf-958101b4bb48-kube-api-access-ckw6t\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.841446 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-catalog-content\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.841687 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-utilities\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.943912 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-utilities\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.945048 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckw6t\" (UniqueName: \"kubernetes.io/projected/a60347a1-afa5-4191-9daf-958101b4bb48-kube-api-access-ckw6t\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.945161 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-catalog-content\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.945643 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-catalog-content\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.944874 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-utilities\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:04 crc kubenswrapper[4848]: I1204 15:36:04.972692 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckw6t\" (UniqueName: \"kubernetes.io/projected/a60347a1-afa5-4191-9daf-958101b4bb48-kube-api-access-ckw6t\") pod \"redhat-marketplace-xfpx5\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:05 crc kubenswrapper[4848]: I1204 15:36:05.029298 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:05 crc kubenswrapper[4848]: I1204 15:36:05.751083 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfpx5"] Dec 04 15:36:06 crc kubenswrapper[4848]: I1204 15:36:06.624893 4848 generic.go:334] "Generic (PLEG): container finished" podID="a60347a1-afa5-4191-9daf-958101b4bb48" containerID="faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2" exitCode=0 Dec 04 15:36:06 crc kubenswrapper[4848]: I1204 15:36:06.624932 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfpx5" event={"ID":"a60347a1-afa5-4191-9daf-958101b4bb48","Type":"ContainerDied","Data":"faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2"} Dec 04 15:36:06 crc kubenswrapper[4848]: I1204 15:36:06.625410 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfpx5" event={"ID":"a60347a1-afa5-4191-9daf-958101b4bb48","Type":"ContainerStarted","Data":"13db13825e72f536f54518627e356698c3978b746aa07f5e7509557c73f93606"} Dec 04 15:36:07 crc kubenswrapper[4848]: I1204 15:36:07.274086 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:36:07 crc kubenswrapper[4848]: I1204 15:36:07.274153 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:36:07 crc kubenswrapper[4848]: I1204 15:36:07.323911 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:36:07 crc kubenswrapper[4848]: I1204 15:36:07.639440 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfpx5" event={"ID":"a60347a1-afa5-4191-9daf-958101b4bb48","Type":"ContainerStarted","Data":"f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51"} Dec 04 15:36:07 crc kubenswrapper[4848]: I1204 15:36:07.690313 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:36:08 crc kubenswrapper[4848]: I1204 15:36:08.650463 4848 generic.go:334] "Generic (PLEG): container finished" podID="a60347a1-afa5-4191-9daf-958101b4bb48" containerID="f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51" exitCode=0 Dec 04 15:36:08 crc kubenswrapper[4848]: I1204 15:36:08.650575 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfpx5" event={"ID":"a60347a1-afa5-4191-9daf-958101b4bb48","Type":"ContainerDied","Data":"f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51"} Dec 04 15:36:09 crc kubenswrapper[4848]: I1204 15:36:09.666084 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfpx5" event={"ID":"a60347a1-afa5-4191-9daf-958101b4bb48","Type":"ContainerStarted","Data":"b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af"} Dec 04 15:36:09 crc kubenswrapper[4848]: I1204 15:36:09.667256 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-974ms"] Dec 04 15:36:09 crc kubenswrapper[4848]: I1204 15:36:09.668121 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-974ms" podUID="78a42305-57db-4198-81b0-4f419c6d4544" containerName="registry-server" containerID="cri-o://eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f" gracePeriod=2 Dec 04 15:36:09 crc kubenswrapper[4848]: I1204 15:36:09.696857 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xfpx5" podStartSLOduration=3.227347836 podStartE2EDuration="5.696836774s" podCreationTimestamp="2025-12-04 15:36:04 +0000 UTC" firstStartedPulling="2025-12-04 15:36:06.626999083 +0000 UTC m=+6470.569495611" lastFinishedPulling="2025-12-04 15:36:09.096488021 +0000 UTC m=+6473.038984549" observedRunningTime="2025-12-04 15:36:09.687919574 +0000 UTC m=+6473.630416102" watchObservedRunningTime="2025-12-04 15:36:09.696836774 +0000 UTC m=+6473.639333302" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.237863 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.379900 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-utilities\") pod \"78a42305-57db-4198-81b0-4f419c6d4544\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.380299 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcgr2\" (UniqueName: \"kubernetes.io/projected/78a42305-57db-4198-81b0-4f419c6d4544-kube-api-access-mcgr2\") pod \"78a42305-57db-4198-81b0-4f419c6d4544\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.380373 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-catalog-content\") pod \"78a42305-57db-4198-81b0-4f419c6d4544\" (UID: \"78a42305-57db-4198-81b0-4f419c6d4544\") " Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.384703 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-utilities" (OuterVolumeSpecName: "utilities") pod "78a42305-57db-4198-81b0-4f419c6d4544" (UID: "78a42305-57db-4198-81b0-4f419c6d4544"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.394574 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78a42305-57db-4198-81b0-4f419c6d4544-kube-api-access-mcgr2" (OuterVolumeSpecName: "kube-api-access-mcgr2") pod "78a42305-57db-4198-81b0-4f419c6d4544" (UID: "78a42305-57db-4198-81b0-4f419c6d4544"). InnerVolumeSpecName "kube-api-access-mcgr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.484155 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcgr2\" (UniqueName: \"kubernetes.io/projected/78a42305-57db-4198-81b0-4f419c6d4544-kube-api-access-mcgr2\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.484200 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.499813 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78a42305-57db-4198-81b0-4f419c6d4544" (UID: "78a42305-57db-4198-81b0-4f419c6d4544"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.586258 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78a42305-57db-4198-81b0-4f419c6d4544-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.681258 4848 generic.go:334] "Generic (PLEG): container finished" podID="78a42305-57db-4198-81b0-4f419c6d4544" containerID="eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f" exitCode=0 Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.681361 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-974ms" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.682294 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-974ms" event={"ID":"78a42305-57db-4198-81b0-4f419c6d4544","Type":"ContainerDied","Data":"eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f"} Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.682323 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-974ms" event={"ID":"78a42305-57db-4198-81b0-4f419c6d4544","Type":"ContainerDied","Data":"ec44ed255d1443438fe089361798c560598b1f91cc55883bb603be629652993d"} Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.682341 4848 scope.go:117] "RemoveContainer" containerID="eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.712062 4848 scope.go:117] "RemoveContainer" containerID="0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.738662 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-974ms"] Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.742914 4848 scope.go:117] "RemoveContainer" containerID="9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.752778 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-974ms"] Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.808193 4848 scope.go:117] "RemoveContainer" containerID="eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f" Dec 04 15:36:10 crc kubenswrapper[4848]: E1204 15:36:10.808574 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f\": container with ID starting with eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f not found: ID does not exist" containerID="eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.808608 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f"} err="failed to get container status \"eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f\": rpc error: code = NotFound desc = could not find container \"eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f\": container with ID starting with eb2aba4017855bcf439bed8adc405e0fb54b2d8eed1619f08b145f6c1b02b35f not found: ID does not exist" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.808629 4848 scope.go:117] "RemoveContainer" containerID="0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c" Dec 04 15:36:10 crc kubenswrapper[4848]: E1204 15:36:10.808898 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c\": container with ID starting with 0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c not found: ID does not exist" containerID="0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.808931 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c"} err="failed to get container status \"0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c\": rpc error: code = NotFound desc = could not find container \"0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c\": container with ID starting with 0501b1db11359e93ae7115c6436c8c504fe211ec36898e8570c68d764538ef9c not found: ID does not exist" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.808970 4848 scope.go:117] "RemoveContainer" containerID="9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d" Dec 04 15:36:10 crc kubenswrapper[4848]: E1204 15:36:10.809707 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d\": container with ID starting with 9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d not found: ID does not exist" containerID="9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d" Dec 04 15:36:10 crc kubenswrapper[4848]: I1204 15:36:10.809731 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d"} err="failed to get container status \"9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d\": rpc error: code = NotFound desc = could not find container \"9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d\": container with ID starting with 9141aa5df9524236d184e74efb4095c385d5c4556cb5dc8ff9b35f07e2cdca4d not found: ID does not exist" Dec 04 15:36:12 crc kubenswrapper[4848]: I1204 15:36:12.408161 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78a42305-57db-4198-81b0-4f419c6d4544" path="/var/lib/kubelet/pods/78a42305-57db-4198-81b0-4f419c6d4544/volumes" Dec 04 15:36:14 crc kubenswrapper[4848]: I1204 15:36:14.314521 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:36:14 crc kubenswrapper[4848]: I1204 15:36:14.314976 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.031091 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.031204 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.082923 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.547584 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c5jdc"] Dec 04 15:36:15 crc kubenswrapper[4848]: E1204 15:36:15.548427 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a42305-57db-4198-81b0-4f419c6d4544" containerName="extract-content" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.548445 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a42305-57db-4198-81b0-4f419c6d4544" containerName="extract-content" Dec 04 15:36:15 crc kubenswrapper[4848]: E1204 15:36:15.548465 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a42305-57db-4198-81b0-4f419c6d4544" containerName="registry-server" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.548474 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a42305-57db-4198-81b0-4f419c6d4544" containerName="registry-server" Dec 04 15:36:15 crc kubenswrapper[4848]: E1204 15:36:15.548502 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a42305-57db-4198-81b0-4f419c6d4544" containerName="extract-utilities" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.548510 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a42305-57db-4198-81b0-4f419c6d4544" containerName="extract-utilities" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.548798 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="78a42305-57db-4198-81b0-4f419c6d4544" containerName="registry-server" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.550810 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.584119 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c5jdc"] Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.604693 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-utilities\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.605025 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-catalog-content\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.605110 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjhjl\" (UniqueName: \"kubernetes.io/projected/b4576359-f419-4d0e-a10a-0877c563c835-kube-api-access-bjhjl\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.707390 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjhjl\" (UniqueName: \"kubernetes.io/projected/b4576359-f419-4d0e-a10a-0877c563c835-kube-api-access-bjhjl\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.707443 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-catalog-content\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.708077 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-utilities\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.708492 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-catalog-content\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.708583 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-utilities\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.739139 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjhjl\" (UniqueName: \"kubernetes.io/projected/b4576359-f419-4d0e-a10a-0877c563c835-kube-api-access-bjhjl\") pod \"community-operators-c5jdc\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.800348 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:15 crc kubenswrapper[4848]: I1204 15:36:15.890507 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:16 crc kubenswrapper[4848]: I1204 15:36:16.415576 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c5jdc"] Dec 04 15:36:16 crc kubenswrapper[4848]: I1204 15:36:16.758743 4848 generic.go:334] "Generic (PLEG): container finished" podID="b4576359-f419-4d0e-a10a-0877c563c835" containerID="de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c" exitCode=0 Dec 04 15:36:16 crc kubenswrapper[4848]: I1204 15:36:16.761104 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5jdc" event={"ID":"b4576359-f419-4d0e-a10a-0877c563c835","Type":"ContainerDied","Data":"de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c"} Dec 04 15:36:16 crc kubenswrapper[4848]: I1204 15:36:16.761170 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5jdc" event={"ID":"b4576359-f419-4d0e-a10a-0877c563c835","Type":"ContainerStarted","Data":"44a8e4f90cac4314c9acecbd348b1ecef775b9570918a7a52cf4c850cdc05faf"} Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.126593 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfpx5"] Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.127397 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xfpx5" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" containerName="registry-server" containerID="cri-o://b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af" gracePeriod=2 Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.770711 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.782789 4848 generic.go:334] "Generic (PLEG): container finished" podID="a60347a1-afa5-4191-9daf-958101b4bb48" containerID="b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af" exitCode=0 Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.782855 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfpx5" event={"ID":"a60347a1-afa5-4191-9daf-958101b4bb48","Type":"ContainerDied","Data":"b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af"} Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.782882 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfpx5" event={"ID":"a60347a1-afa5-4191-9daf-958101b4bb48","Type":"ContainerDied","Data":"13db13825e72f536f54518627e356698c3978b746aa07f5e7509557c73f93606"} Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.782900 4848 scope.go:117] "RemoveContainer" containerID="b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.783062 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfpx5" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.787091 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5jdc" event={"ID":"b4576359-f419-4d0e-a10a-0877c563c835","Type":"ContainerStarted","Data":"997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4"} Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.810332 4848 scope.go:117] "RemoveContainer" containerID="f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.843875 4848 scope.go:117] "RemoveContainer" containerID="faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.902402 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-utilities\") pod \"a60347a1-afa5-4191-9daf-958101b4bb48\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.902462 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckw6t\" (UniqueName: \"kubernetes.io/projected/a60347a1-afa5-4191-9daf-958101b4bb48-kube-api-access-ckw6t\") pod \"a60347a1-afa5-4191-9daf-958101b4bb48\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.902672 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-catalog-content\") pod \"a60347a1-afa5-4191-9daf-958101b4bb48\" (UID: \"a60347a1-afa5-4191-9daf-958101b4bb48\") " Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.903480 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-utilities" (OuterVolumeSpecName: "utilities") pod "a60347a1-afa5-4191-9daf-958101b4bb48" (UID: "a60347a1-afa5-4191-9daf-958101b4bb48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.903590 4848 scope.go:117] "RemoveContainer" containerID="b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af" Dec 04 15:36:18 crc kubenswrapper[4848]: E1204 15:36:18.905870 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af\": container with ID starting with b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af not found: ID does not exist" containerID="b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.905922 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af"} err="failed to get container status \"b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af\": rpc error: code = NotFound desc = could not find container \"b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af\": container with ID starting with b0a926f4636ac7402de63237c19289496f6341a52a10496f2610c83c75c0d4af not found: ID does not exist" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.905968 4848 scope.go:117] "RemoveContainer" containerID="f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51" Dec 04 15:36:18 crc kubenswrapper[4848]: E1204 15:36:18.906527 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51\": container with ID starting with f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51 not found: ID does not exist" containerID="f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.906599 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51"} err="failed to get container status \"f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51\": rpc error: code = NotFound desc = could not find container \"f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51\": container with ID starting with f78c6b502378d8405ecd53e998ba025b91b086f91ddfe40ef3cd3c93005d0a51 not found: ID does not exist" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.906625 4848 scope.go:117] "RemoveContainer" containerID="faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2" Dec 04 15:36:18 crc kubenswrapper[4848]: E1204 15:36:18.907101 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2\": container with ID starting with faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2 not found: ID does not exist" containerID="faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.907142 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2"} err="failed to get container status \"faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2\": rpc error: code = NotFound desc = could not find container \"faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2\": container with ID starting with faa4b36166b998b32bbfe659f34c815d7f8473dfc55ef2c1b345fcafdaf4eea2 not found: ID does not exist" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.908633 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a60347a1-afa5-4191-9daf-958101b4bb48-kube-api-access-ckw6t" (OuterVolumeSpecName: "kube-api-access-ckw6t") pod "a60347a1-afa5-4191-9daf-958101b4bb48" (UID: "a60347a1-afa5-4191-9daf-958101b4bb48"). InnerVolumeSpecName "kube-api-access-ckw6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:36:18 crc kubenswrapper[4848]: I1204 15:36:18.921440 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a60347a1-afa5-4191-9daf-958101b4bb48" (UID: "a60347a1-afa5-4191-9daf-958101b4bb48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:36:19 crc kubenswrapper[4848]: I1204 15:36:19.005989 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:19 crc kubenswrapper[4848]: I1204 15:36:19.006046 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a60347a1-afa5-4191-9daf-958101b4bb48-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:19 crc kubenswrapper[4848]: I1204 15:36:19.006059 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckw6t\" (UniqueName: \"kubernetes.io/projected/a60347a1-afa5-4191-9daf-958101b4bb48-kube-api-access-ckw6t\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:19 crc kubenswrapper[4848]: I1204 15:36:19.186917 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfpx5"] Dec 04 15:36:19 crc kubenswrapper[4848]: I1204 15:36:19.199650 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfpx5"] Dec 04 15:36:19 crc kubenswrapper[4848]: I1204 15:36:19.801649 4848 generic.go:334] "Generic (PLEG): container finished" podID="b4576359-f419-4d0e-a10a-0877c563c835" containerID="997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4" exitCode=0 Dec 04 15:36:19 crc kubenswrapper[4848]: I1204 15:36:19.801691 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5jdc" event={"ID":"b4576359-f419-4d0e-a10a-0877c563c835","Type":"ContainerDied","Data":"997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4"} Dec 04 15:36:20 crc kubenswrapper[4848]: I1204 15:36:20.408665 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" path="/var/lib/kubelet/pods/a60347a1-afa5-4191-9daf-958101b4bb48/volumes" Dec 04 15:36:20 crc kubenswrapper[4848]: I1204 15:36:20.814492 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5jdc" event={"ID":"b4576359-f419-4d0e-a10a-0877c563c835","Type":"ContainerStarted","Data":"05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae"} Dec 04 15:36:20 crc kubenswrapper[4848]: I1204 15:36:20.841751 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c5jdc" podStartSLOduration=2.411881302 podStartE2EDuration="5.841731355s" podCreationTimestamp="2025-12-04 15:36:15 +0000 UTC" firstStartedPulling="2025-12-04 15:36:16.762301813 +0000 UTC m=+6480.704798341" lastFinishedPulling="2025-12-04 15:36:20.192151866 +0000 UTC m=+6484.134648394" observedRunningTime="2025-12-04 15:36:20.830431147 +0000 UTC m=+6484.772927675" watchObservedRunningTime="2025-12-04 15:36:20.841731355 +0000 UTC m=+6484.784227883" Dec 04 15:36:25 crc kubenswrapper[4848]: I1204 15:36:25.891589 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:25 crc kubenswrapper[4848]: I1204 15:36:25.893497 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:25 crc kubenswrapper[4848]: I1204 15:36:25.942445 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:26 crc kubenswrapper[4848]: I1204 15:36:26.930930 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:27 crc kubenswrapper[4848]: I1204 15:36:27.186387 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c5jdc"] Dec 04 15:36:28 crc kubenswrapper[4848]: I1204 15:36:28.890657 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c5jdc" podUID="b4576359-f419-4d0e-a10a-0877c563c835" containerName="registry-server" containerID="cri-o://05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae" gracePeriod=2 Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.428828 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.467832 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjhjl\" (UniqueName: \"kubernetes.io/projected/b4576359-f419-4d0e-a10a-0877c563c835-kube-api-access-bjhjl\") pod \"b4576359-f419-4d0e-a10a-0877c563c835\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.468157 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-utilities\") pod \"b4576359-f419-4d0e-a10a-0877c563c835\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.468377 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-catalog-content\") pod \"b4576359-f419-4d0e-a10a-0877c563c835\" (UID: \"b4576359-f419-4d0e-a10a-0877c563c835\") " Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.487051 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-utilities" (OuterVolumeSpecName: "utilities") pod "b4576359-f419-4d0e-a10a-0877c563c835" (UID: "b4576359-f419-4d0e-a10a-0877c563c835"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.525266 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4576359-f419-4d0e-a10a-0877c563c835-kube-api-access-bjhjl" (OuterVolumeSpecName: "kube-api-access-bjhjl") pod "b4576359-f419-4d0e-a10a-0877c563c835" (UID: "b4576359-f419-4d0e-a10a-0877c563c835"). InnerVolumeSpecName "kube-api-access-bjhjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.571974 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.572014 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjhjl\" (UniqueName: \"kubernetes.io/projected/b4576359-f419-4d0e-a10a-0877c563c835-kube-api-access-bjhjl\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.597667 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4576359-f419-4d0e-a10a-0877c563c835" (UID: "b4576359-f419-4d0e-a10a-0877c563c835"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.673673 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4576359-f419-4d0e-a10a-0877c563c835-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.905280 4848 generic.go:334] "Generic (PLEG): container finished" podID="b4576359-f419-4d0e-a10a-0877c563c835" containerID="05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae" exitCode=0 Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.905326 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5jdc" event={"ID":"b4576359-f419-4d0e-a10a-0877c563c835","Type":"ContainerDied","Data":"05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae"} Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.905359 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5jdc" event={"ID":"b4576359-f419-4d0e-a10a-0877c563c835","Type":"ContainerDied","Data":"44a8e4f90cac4314c9acecbd348b1ecef775b9570918a7a52cf4c850cdc05faf"} Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.905376 4848 scope.go:117] "RemoveContainer" containerID="05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.905382 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5jdc" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.929529 4848 scope.go:117] "RemoveContainer" containerID="997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4" Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.959089 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c5jdc"] Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.971551 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c5jdc"] Dec 04 15:36:29 crc kubenswrapper[4848]: I1204 15:36:29.990816 4848 scope.go:117] "RemoveContainer" containerID="de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c" Dec 04 15:36:30 crc kubenswrapper[4848]: I1204 15:36:30.025420 4848 scope.go:117] "RemoveContainer" containerID="05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae" Dec 04 15:36:30 crc kubenswrapper[4848]: E1204 15:36:30.026216 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae\": container with ID starting with 05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae not found: ID does not exist" containerID="05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae" Dec 04 15:36:30 crc kubenswrapper[4848]: I1204 15:36:30.026251 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae"} err="failed to get container status \"05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae\": rpc error: code = NotFound desc = could not find container \"05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae\": container with ID starting with 05216ffe4b0c3af834fad1d535d39cc3b4914e511603116d9dbb06c1143f1aae not found: ID does not exist" Dec 04 15:36:30 crc kubenswrapper[4848]: I1204 15:36:30.026278 4848 scope.go:117] "RemoveContainer" containerID="997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4" Dec 04 15:36:30 crc kubenswrapper[4848]: E1204 15:36:30.026609 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4\": container with ID starting with 997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4 not found: ID does not exist" containerID="997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4" Dec 04 15:36:30 crc kubenswrapper[4848]: I1204 15:36:30.026649 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4"} err="failed to get container status \"997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4\": rpc error: code = NotFound desc = could not find container \"997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4\": container with ID starting with 997dd97a8d338fabf46ced6c410a66b9ee4a5de63e757fdc8e222cc33cf653f4 not found: ID does not exist" Dec 04 15:36:30 crc kubenswrapper[4848]: I1204 15:36:30.026677 4848 scope.go:117] "RemoveContainer" containerID="de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c" Dec 04 15:36:30 crc kubenswrapper[4848]: E1204 15:36:30.027017 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c\": container with ID starting with de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c not found: ID does not exist" containerID="de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c" Dec 04 15:36:30 crc kubenswrapper[4848]: I1204 15:36:30.027327 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c"} err="failed to get container status \"de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c\": rpc error: code = NotFound desc = could not find container \"de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c\": container with ID starting with de52d4ec0d43542d010fc64ff9e2270a445fb3705fd17c0b00bf5a87d8a0d13c not found: ID does not exist" Dec 04 15:36:30 crc kubenswrapper[4848]: I1204 15:36:30.408713 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4576359-f419-4d0e-a10a-0877c563c835" path="/var/lib/kubelet/pods/b4576359-f419-4d0e-a10a-0877c563c835/volumes" Dec 04 15:36:42 crc kubenswrapper[4848]: I1204 15:36:42.440569 4848 scope.go:117] "RemoveContainer" containerID="6395027a8d8f3bedf2d4dc2b1c1a115f824419d72acdd7f2c1a2a0befb6059aa" Dec 04 15:36:44 crc kubenswrapper[4848]: I1204 15:36:44.314873 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:36:44 crc kubenswrapper[4848]: I1204 15:36:44.315608 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:36:44 crc kubenswrapper[4848]: I1204 15:36:44.315689 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:36:44 crc kubenswrapper[4848]: I1204 15:36:44.318175 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:36:44 crc kubenswrapper[4848]: I1204 15:36:44.318313 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" gracePeriod=600 Dec 04 15:36:44 crc kubenswrapper[4848]: E1204 15:36:44.581838 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:36:45 crc kubenswrapper[4848]: I1204 15:36:45.072772 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" exitCode=0 Dec 04 15:36:45 crc kubenswrapper[4848]: I1204 15:36:45.072847 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760"} Dec 04 15:36:45 crc kubenswrapper[4848]: I1204 15:36:45.073246 4848 scope.go:117] "RemoveContainer" containerID="6e8b1d4f65c5c25aa19d0b9bebebf377fde86d7d15e1e29fa678a12b8a93b1e9" Dec 04 15:36:45 crc kubenswrapper[4848]: I1204 15:36:45.074002 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:36:45 crc kubenswrapper[4848]: E1204 15:36:45.074473 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:36:58 crc kubenswrapper[4848]: I1204 15:36:58.393548 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:36:58 crc kubenswrapper[4848]: E1204 15:36:58.394249 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:37:09 crc kubenswrapper[4848]: I1204 15:37:09.395022 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:37:09 crc kubenswrapper[4848]: E1204 15:37:09.395822 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:37:10 crc kubenswrapper[4848]: I1204 15:37:10.367203 4848 generic.go:334] "Generic (PLEG): container finished" podID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerID="579b3ae0082fabe61435e5fcf9bc6d06ca43a91cea8ea591c17c04042fa131d4" exitCode=0 Dec 04 15:37:10 crc kubenswrapper[4848]: I1204 15:37:10.367284 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m2chb/must-gather-pjmd4" event={"ID":"38bd2da4-bb47-492f-9bd3-fd64625a0ceb","Type":"ContainerDied","Data":"579b3ae0082fabe61435e5fcf9bc6d06ca43a91cea8ea591c17c04042fa131d4"} Dec 04 15:37:10 crc kubenswrapper[4848]: I1204 15:37:10.369225 4848 scope.go:117] "RemoveContainer" containerID="579b3ae0082fabe61435e5fcf9bc6d06ca43a91cea8ea591c17c04042fa131d4" Dec 04 15:37:10 crc kubenswrapper[4848]: I1204 15:37:10.963399 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m2chb_must-gather-pjmd4_38bd2da4-bb47-492f-9bd3-fd64625a0ceb/gather/0.log" Dec 04 15:37:19 crc kubenswrapper[4848]: I1204 15:37:19.027650 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m2chb/must-gather-pjmd4"] Dec 04 15:37:19 crc kubenswrapper[4848]: I1204 15:37:19.028614 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-m2chb/must-gather-pjmd4" podUID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerName="copy" containerID="cri-o://edb90f767b4c56392ebbaa205266253d073cb5cdfac6b23d7036a3c08a10b1bc" gracePeriod=2 Dec 04 15:37:19 crc kubenswrapper[4848]: I1204 15:37:19.046800 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m2chb/must-gather-pjmd4"] Dec 04 15:37:19 crc kubenswrapper[4848]: I1204 15:37:19.483215 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m2chb_must-gather-pjmd4_38bd2da4-bb47-492f-9bd3-fd64625a0ceb/copy/0.log" Dec 04 15:37:19 crc kubenswrapper[4848]: I1204 15:37:19.483911 4848 generic.go:334] "Generic (PLEG): container finished" podID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerID="edb90f767b4c56392ebbaa205266253d073cb5cdfac6b23d7036a3c08a10b1bc" exitCode=143 Dec 04 15:37:19 crc kubenswrapper[4848]: E1204 15:37:19.488394 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38bd2da4_bb47_492f_9bd3_fd64625a0ceb.slice/crio-conmon-edb90f767b4c56392ebbaa205266253d073cb5cdfac6b23d7036a3c08a10b1bc.scope\": RecentStats: unable to find data in memory cache]" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.094051 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m2chb_must-gather-pjmd4_38bd2da4-bb47-492f-9bd3-fd64625a0ceb/copy/0.log" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.094773 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.208413 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-must-gather-output\") pod \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\" (UID: \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\") " Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.208477 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vplrz\" (UniqueName: \"kubernetes.io/projected/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-kube-api-access-vplrz\") pod \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\" (UID: \"38bd2da4-bb47-492f-9bd3-fd64625a0ceb\") " Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.215348 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-kube-api-access-vplrz" (OuterVolumeSpecName: "kube-api-access-vplrz") pod "38bd2da4-bb47-492f-9bd3-fd64625a0ceb" (UID: "38bd2da4-bb47-492f-9bd3-fd64625a0ceb"). InnerVolumeSpecName "kube-api-access-vplrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.311752 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vplrz\" (UniqueName: \"kubernetes.io/projected/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-kube-api-access-vplrz\") on node \"crc\" DevicePath \"\"" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.387530 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "38bd2da4-bb47-492f-9bd3-fd64625a0ceb" (UID: "38bd2da4-bb47-492f-9bd3-fd64625a0ceb"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.406777 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" path="/var/lib/kubelet/pods/38bd2da4-bb47-492f-9bd3-fd64625a0ceb/volumes" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.413602 4848 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bd2da4-bb47-492f-9bd3-fd64625a0ceb-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.496216 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m2chb_must-gather-pjmd4_38bd2da4-bb47-492f-9bd3-fd64625a0ceb/copy/0.log" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.496688 4848 scope.go:117] "RemoveContainer" containerID="edb90f767b4c56392ebbaa205266253d073cb5cdfac6b23d7036a3c08a10b1bc" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.496765 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m2chb/must-gather-pjmd4" Dec 04 15:37:20 crc kubenswrapper[4848]: I1204 15:37:20.543650 4848 scope.go:117] "RemoveContainer" containerID="579b3ae0082fabe61435e5fcf9bc6d06ca43a91cea8ea591c17c04042fa131d4" Dec 04 15:37:22 crc kubenswrapper[4848]: I1204 15:37:22.394918 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:37:22 crc kubenswrapper[4848]: E1204 15:37:22.395842 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:37:37 crc kubenswrapper[4848]: I1204 15:37:37.394504 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:37:37 crc kubenswrapper[4848]: E1204 15:37:37.395329 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:37:42 crc kubenswrapper[4848]: I1204 15:37:42.584185 4848 scope.go:117] "RemoveContainer" containerID="3bd5ef924b389175f831c49b9f495ae6ef9a28ddd9930b6a7ef0b74cd47303d5" Dec 04 15:37:52 crc kubenswrapper[4848]: I1204 15:37:52.393702 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:37:52 crc kubenswrapper[4848]: E1204 15:37:52.394542 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:38:03 crc kubenswrapper[4848]: I1204 15:38:03.393487 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:38:03 crc kubenswrapper[4848]: E1204 15:38:03.394431 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:38:15 crc kubenswrapper[4848]: I1204 15:38:15.393926 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:38:15 crc kubenswrapper[4848]: E1204 15:38:15.394895 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:38:29 crc kubenswrapper[4848]: I1204 15:38:29.393915 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:38:29 crc kubenswrapper[4848]: E1204 15:38:29.394820 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:38:44 crc kubenswrapper[4848]: I1204 15:38:44.393790 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:38:44 crc kubenswrapper[4848]: E1204 15:38:44.394654 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:38:57 crc kubenswrapper[4848]: I1204 15:38:57.393456 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:38:57 crc kubenswrapper[4848]: E1204 15:38:57.394350 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:39:12 crc kubenswrapper[4848]: I1204 15:39:12.394022 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:39:12 crc kubenswrapper[4848]: E1204 15:39:12.395244 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:39:24 crc kubenswrapper[4848]: I1204 15:39:24.394254 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:39:24 crc kubenswrapper[4848]: E1204 15:39:24.396168 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:39:37 crc kubenswrapper[4848]: I1204 15:39:37.394192 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:39:37 crc kubenswrapper[4848]: E1204 15:39:37.395246 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.435148 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zcwln"] Dec 04 15:39:42 crc kubenswrapper[4848]: E1204 15:39:42.436164 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" containerName="extract-utilities" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436183 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" containerName="extract-utilities" Dec 04 15:39:42 crc kubenswrapper[4848]: E1204 15:39:42.436204 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" containerName="extract-content" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436211 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" containerName="extract-content" Dec 04 15:39:42 crc kubenswrapper[4848]: E1204 15:39:42.436220 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4576359-f419-4d0e-a10a-0877c563c835" containerName="extract-content" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436227 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4576359-f419-4d0e-a10a-0877c563c835" containerName="extract-content" Dec 04 15:39:42 crc kubenswrapper[4848]: E1204 15:39:42.436258 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerName="copy" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436266 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerName="copy" Dec 04 15:39:42 crc kubenswrapper[4848]: E1204 15:39:42.436287 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4576359-f419-4d0e-a10a-0877c563c835" containerName="extract-utilities" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436295 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4576359-f419-4d0e-a10a-0877c563c835" containerName="extract-utilities" Dec 04 15:39:42 crc kubenswrapper[4848]: E1204 15:39:42.436324 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" containerName="registry-server" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436330 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" containerName="registry-server" Dec 04 15:39:42 crc kubenswrapper[4848]: E1204 15:39:42.436342 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4576359-f419-4d0e-a10a-0877c563c835" containerName="registry-server" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436348 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4576359-f419-4d0e-a10a-0877c563c835" containerName="registry-server" Dec 04 15:39:42 crc kubenswrapper[4848]: E1204 15:39:42.436358 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerName="gather" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436364 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerName="gather" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436587 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerName="copy" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436610 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4576359-f419-4d0e-a10a-0877c563c835" containerName="registry-server" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436637 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="38bd2da4-bb47-492f-9bd3-fd64625a0ceb" containerName="gather" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.436647 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a60347a1-afa5-4191-9daf-958101b4bb48" containerName="registry-server" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.438409 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.449034 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zcwln"] Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.588722 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-catalog-content\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.588876 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8z44\" (UniqueName: \"kubernetes.io/projected/a986ee05-07bc-43f2-96e1-45e55b31ac20-kube-api-access-z8z44\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.588937 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-utilities\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.691299 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-catalog-content\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.691404 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8z44\" (UniqueName: \"kubernetes.io/projected/a986ee05-07bc-43f2-96e1-45e55b31ac20-kube-api-access-z8z44\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.691457 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-utilities\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.692239 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-utilities\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.692250 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-catalog-content\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.711392 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8z44\" (UniqueName: \"kubernetes.io/projected/a986ee05-07bc-43f2-96e1-45e55b31ac20-kube-api-access-z8z44\") pod \"redhat-operators-zcwln\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:42 crc kubenswrapper[4848]: I1204 15:39:42.761738 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:43 crc kubenswrapper[4848]: I1204 15:39:43.217329 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zcwln"] Dec 04 15:39:44 crc kubenswrapper[4848]: I1204 15:39:44.065446 4848 generic.go:334] "Generic (PLEG): container finished" podID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerID="3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39" exitCode=0 Dec 04 15:39:44 crc kubenswrapper[4848]: I1204 15:39:44.065509 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcwln" event={"ID":"a986ee05-07bc-43f2-96e1-45e55b31ac20","Type":"ContainerDied","Data":"3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39"} Dec 04 15:39:44 crc kubenswrapper[4848]: I1204 15:39:44.065786 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcwln" event={"ID":"a986ee05-07bc-43f2-96e1-45e55b31ac20","Type":"ContainerStarted","Data":"33448847ad23026c4215f02a44e32afdce2e947e25a4bdf4d04ec4e898562fa9"} Dec 04 15:39:45 crc kubenswrapper[4848]: I1204 15:39:45.077396 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcwln" event={"ID":"a986ee05-07bc-43f2-96e1-45e55b31ac20","Type":"ContainerStarted","Data":"cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578"} Dec 04 15:39:49 crc kubenswrapper[4848]: I1204 15:39:49.126957 4848 generic.go:334] "Generic (PLEG): container finished" podID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerID="cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578" exitCode=0 Dec 04 15:39:49 crc kubenswrapper[4848]: I1204 15:39:49.127017 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcwln" event={"ID":"a986ee05-07bc-43f2-96e1-45e55b31ac20","Type":"ContainerDied","Data":"cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578"} Dec 04 15:39:50 crc kubenswrapper[4848]: I1204 15:39:50.143220 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcwln" event={"ID":"a986ee05-07bc-43f2-96e1-45e55b31ac20","Type":"ContainerStarted","Data":"7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a"} Dec 04 15:39:50 crc kubenswrapper[4848]: I1204 15:39:50.176213 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zcwln" podStartSLOduration=2.704224265 podStartE2EDuration="8.176190373s" podCreationTimestamp="2025-12-04 15:39:42 +0000 UTC" firstStartedPulling="2025-12-04 15:39:44.068366342 +0000 UTC m=+6688.010862860" lastFinishedPulling="2025-12-04 15:39:49.54033244 +0000 UTC m=+6693.482828968" observedRunningTime="2025-12-04 15:39:50.167463749 +0000 UTC m=+6694.109960277" watchObservedRunningTime="2025-12-04 15:39:50.176190373 +0000 UTC m=+6694.118686901" Dec 04 15:39:52 crc kubenswrapper[4848]: I1204 15:39:52.393721 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:39:52 crc kubenswrapper[4848]: E1204 15:39:52.394563 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:39:52 crc kubenswrapper[4848]: I1204 15:39:52.762291 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:52 crc kubenswrapper[4848]: I1204 15:39:52.762330 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:39:53 crc kubenswrapper[4848]: I1204 15:39:53.812455 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zcwln" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="registry-server" probeResult="failure" output=< Dec 04 15:39:53 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:39:53 crc kubenswrapper[4848]: > Dec 04 15:40:02 crc kubenswrapper[4848]: I1204 15:40:02.815161 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:40:02 crc kubenswrapper[4848]: I1204 15:40:02.874643 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:40:03 crc kubenswrapper[4848]: I1204 15:40:03.057588 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zcwln"] Dec 04 15:40:03 crc kubenswrapper[4848]: I1204 15:40:03.393503 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:40:03 crc kubenswrapper[4848]: E1204 15:40:03.393846 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:40:04 crc kubenswrapper[4848]: I1204 15:40:04.292744 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zcwln" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="registry-server" containerID="cri-o://7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a" gracePeriod=2 Dec 04 15:40:04 crc kubenswrapper[4848]: I1204 15:40:04.821250 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:40:04 crc kubenswrapper[4848]: I1204 15:40:04.936746 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-utilities\") pod \"a986ee05-07bc-43f2-96e1-45e55b31ac20\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " Dec 04 15:40:04 crc kubenswrapper[4848]: I1204 15:40:04.937143 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8z44\" (UniqueName: \"kubernetes.io/projected/a986ee05-07bc-43f2-96e1-45e55b31ac20-kube-api-access-z8z44\") pod \"a986ee05-07bc-43f2-96e1-45e55b31ac20\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " Dec 04 15:40:04 crc kubenswrapper[4848]: I1204 15:40:04.937442 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-catalog-content\") pod \"a986ee05-07bc-43f2-96e1-45e55b31ac20\" (UID: \"a986ee05-07bc-43f2-96e1-45e55b31ac20\") " Dec 04 15:40:04 crc kubenswrapper[4848]: I1204 15:40:04.937735 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-utilities" (OuterVolumeSpecName: "utilities") pod "a986ee05-07bc-43f2-96e1-45e55b31ac20" (UID: "a986ee05-07bc-43f2-96e1-45e55b31ac20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:40:04 crc kubenswrapper[4848]: I1204 15:40:04.937936 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:40:04 crc kubenswrapper[4848]: I1204 15:40:04.945117 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a986ee05-07bc-43f2-96e1-45e55b31ac20-kube-api-access-z8z44" (OuterVolumeSpecName: "kube-api-access-z8z44") pod "a986ee05-07bc-43f2-96e1-45e55b31ac20" (UID: "a986ee05-07bc-43f2-96e1-45e55b31ac20"). InnerVolumeSpecName "kube-api-access-z8z44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.040763 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8z44\" (UniqueName: \"kubernetes.io/projected/a986ee05-07bc-43f2-96e1-45e55b31ac20-kube-api-access-z8z44\") on node \"crc\" DevicePath \"\"" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.053914 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a986ee05-07bc-43f2-96e1-45e55b31ac20" (UID: "a986ee05-07bc-43f2-96e1-45e55b31ac20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.142226 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986ee05-07bc-43f2-96e1-45e55b31ac20-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.308443 4848 generic.go:334] "Generic (PLEG): container finished" podID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerID="7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a" exitCode=0 Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.308492 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcwln" event={"ID":"a986ee05-07bc-43f2-96e1-45e55b31ac20","Type":"ContainerDied","Data":"7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a"} Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.308551 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zcwln" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.308582 4848 scope.go:117] "RemoveContainer" containerID="7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.308564 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zcwln" event={"ID":"a986ee05-07bc-43f2-96e1-45e55b31ac20","Type":"ContainerDied","Data":"33448847ad23026c4215f02a44e32afdce2e947e25a4bdf4d04ec4e898562fa9"} Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.329411 4848 scope.go:117] "RemoveContainer" containerID="cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.357723 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zcwln"] Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.363158 4848 scope.go:117] "RemoveContainer" containerID="3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.371087 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zcwln"] Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.418639 4848 scope.go:117] "RemoveContainer" containerID="7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a" Dec 04 15:40:05 crc kubenswrapper[4848]: E1204 15:40:05.419226 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a\": container with ID starting with 7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a not found: ID does not exist" containerID="7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.419274 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a"} err="failed to get container status \"7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a\": rpc error: code = NotFound desc = could not find container \"7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a\": container with ID starting with 7456ddb3ac9cc3f1da8e7f86b013cb14b9b2c6cd96ee22c41134b97e1688fe9a not found: ID does not exist" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.419302 4848 scope.go:117] "RemoveContainer" containerID="cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578" Dec 04 15:40:05 crc kubenswrapper[4848]: E1204 15:40:05.419853 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578\": container with ID starting with cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578 not found: ID does not exist" containerID="cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.419894 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578"} err="failed to get container status \"cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578\": rpc error: code = NotFound desc = could not find container \"cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578\": container with ID starting with cf20a31a7653f9b07cd0e7934d86babdd9a3f6a92da1abe87da4c4395707a578 not found: ID does not exist" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.419922 4848 scope.go:117] "RemoveContainer" containerID="3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39" Dec 04 15:40:05 crc kubenswrapper[4848]: E1204 15:40:05.420252 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39\": container with ID starting with 3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39 not found: ID does not exist" containerID="3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39" Dec 04 15:40:05 crc kubenswrapper[4848]: I1204 15:40:05.420279 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39"} err="failed to get container status \"3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39\": rpc error: code = NotFound desc = could not find container \"3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39\": container with ID starting with 3f4c57e54f23fb3a1563bd45c9423ba3de7730b8fd2f219b240f27bf37178a39 not found: ID does not exist" Dec 04 15:40:06 crc kubenswrapper[4848]: I1204 15:40:06.408438 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" path="/var/lib/kubelet/pods/a986ee05-07bc-43f2-96e1-45e55b31ac20/volumes" Dec 04 15:40:16 crc kubenswrapper[4848]: I1204 15:40:16.404678 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:40:16 crc kubenswrapper[4848]: E1204 15:40:16.405379 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:40:27 crc kubenswrapper[4848]: I1204 15:40:27.393765 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:40:27 crc kubenswrapper[4848]: E1204 15:40:27.394557 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.432011 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9k646/must-gather-vx45g"] Dec 04 15:40:39 crc kubenswrapper[4848]: E1204 15:40:39.433126 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="extract-utilities" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.433144 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="extract-utilities" Dec 04 15:40:39 crc kubenswrapper[4848]: E1204 15:40:39.433207 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="extract-content" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.433216 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="extract-content" Dec 04 15:40:39 crc kubenswrapper[4848]: E1204 15:40:39.433238 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="registry-server" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.433246 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="registry-server" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.433533 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="a986ee05-07bc-43f2-96e1-45e55b31ac20" containerName="registry-server" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.442180 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.448240 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9k646/must-gather-vx45g"] Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.448346 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9k646"/"kube-root-ca.crt" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.448425 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-9k646"/"default-dockercfg-drvb5" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.448574 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9k646"/"openshift-service-ca.crt" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.489997 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf13c537-64dd-4e6b-8de6-9181b448ebd2-must-gather-output\") pod \"must-gather-vx45g\" (UID: \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\") " pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.490054 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dz5c\" (UniqueName: \"kubernetes.io/projected/cf13c537-64dd-4e6b-8de6-9181b448ebd2-kube-api-access-9dz5c\") pod \"must-gather-vx45g\" (UID: \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\") " pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.593039 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf13c537-64dd-4e6b-8de6-9181b448ebd2-must-gather-output\") pod \"must-gather-vx45g\" (UID: \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\") " pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.593106 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dz5c\" (UniqueName: \"kubernetes.io/projected/cf13c537-64dd-4e6b-8de6-9181b448ebd2-kube-api-access-9dz5c\") pod \"must-gather-vx45g\" (UID: \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\") " pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.593545 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf13c537-64dd-4e6b-8de6-9181b448ebd2-must-gather-output\") pod \"must-gather-vx45g\" (UID: \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\") " pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:40:39 crc kubenswrapper[4848]: I1204 15:40:39.622625 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dz5c\" (UniqueName: \"kubernetes.io/projected/cf13c537-64dd-4e6b-8de6-9181b448ebd2-kube-api-access-9dz5c\") pod \"must-gather-vx45g\" (UID: \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\") " pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:40:40 crc kubenswrapper[4848]: I1204 15:40:39.763479 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:40:40 crc kubenswrapper[4848]: I1204 15:40:40.934564 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9k646/must-gather-vx45g"] Dec 04 15:40:41 crc kubenswrapper[4848]: I1204 15:40:41.696654 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/must-gather-vx45g" event={"ID":"cf13c537-64dd-4e6b-8de6-9181b448ebd2","Type":"ContainerStarted","Data":"776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987"} Dec 04 15:40:41 crc kubenswrapper[4848]: I1204 15:40:41.697250 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/must-gather-vx45g" event={"ID":"cf13c537-64dd-4e6b-8de6-9181b448ebd2","Type":"ContainerStarted","Data":"ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e"} Dec 04 15:40:41 crc kubenswrapper[4848]: I1204 15:40:41.697361 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/must-gather-vx45g" event={"ID":"cf13c537-64dd-4e6b-8de6-9181b448ebd2","Type":"ContainerStarted","Data":"fac706525ed8fd894682b8a33fc39f66fb29bd6ac39b55a5bc7ab60f5adb738f"} Dec 04 15:40:41 crc kubenswrapper[4848]: I1204 15:40:41.731191 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9k646/must-gather-vx45g" podStartSLOduration=2.731162481 podStartE2EDuration="2.731162481s" podCreationTimestamp="2025-12-04 15:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 15:40:41.714365209 +0000 UTC m=+6745.656861757" watchObservedRunningTime="2025-12-04 15:40:41.731162481 +0000 UTC m=+6745.673659029" Dec 04 15:40:42 crc kubenswrapper[4848]: I1204 15:40:42.394047 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:40:42 crc kubenswrapper[4848]: E1204 15:40:42.394806 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:40:44 crc kubenswrapper[4848]: I1204 15:40:44.939002 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9k646/crc-debug-tq475"] Dec 04 15:40:44 crc kubenswrapper[4848]: I1204 15:40:44.941241 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.027547 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0c931889-7884-40ab-a9f4-8b90e49fa8e6-host\") pod \"crc-debug-tq475\" (UID: \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\") " pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.027631 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d85w\" (UniqueName: \"kubernetes.io/projected/0c931889-7884-40ab-a9f4-8b90e49fa8e6-kube-api-access-2d85w\") pod \"crc-debug-tq475\" (UID: \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\") " pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.130617 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d85w\" (UniqueName: \"kubernetes.io/projected/0c931889-7884-40ab-a9f4-8b90e49fa8e6-kube-api-access-2d85w\") pod \"crc-debug-tq475\" (UID: \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\") " pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.131321 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0c931889-7884-40ab-a9f4-8b90e49fa8e6-host\") pod \"crc-debug-tq475\" (UID: \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\") " pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.131764 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0c931889-7884-40ab-a9f4-8b90e49fa8e6-host\") pod \"crc-debug-tq475\" (UID: \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\") " pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.152380 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d85w\" (UniqueName: \"kubernetes.io/projected/0c931889-7884-40ab-a9f4-8b90e49fa8e6-kube-api-access-2d85w\") pod \"crc-debug-tq475\" (UID: \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\") " pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.266027 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:40:45 crc kubenswrapper[4848]: W1204 15:40:45.329567 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c931889_7884_40ab_a9f4_8b90e49fa8e6.slice/crio-d7498c5d882be639d8c81ed6cfcdeca30dbd1740b3ce23119c4339318a6f6f20 WatchSource:0}: Error finding container d7498c5d882be639d8c81ed6cfcdeca30dbd1740b3ce23119c4339318a6f6f20: Status 404 returned error can't find the container with id d7498c5d882be639d8c81ed6cfcdeca30dbd1740b3ce23119c4339318a6f6f20 Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.742459 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/crc-debug-tq475" event={"ID":"0c931889-7884-40ab-a9f4-8b90e49fa8e6","Type":"ContainerStarted","Data":"16adfe1c2595a3dea3f34e4665f00ffa325bbc3cf5d1fc04e3618212fd17d281"} Dec 04 15:40:45 crc kubenswrapper[4848]: I1204 15:40:45.743015 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/crc-debug-tq475" event={"ID":"0c931889-7884-40ab-a9f4-8b90e49fa8e6","Type":"ContainerStarted","Data":"d7498c5d882be639d8c81ed6cfcdeca30dbd1740b3ce23119c4339318a6f6f20"} Dec 04 15:40:55 crc kubenswrapper[4848]: I1204 15:40:55.393927 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:40:55 crc kubenswrapper[4848]: E1204 15:40:55.394646 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:41:06 crc kubenswrapper[4848]: I1204 15:41:06.400846 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:41:06 crc kubenswrapper[4848]: E1204 15:41:06.401539 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:41:21 crc kubenswrapper[4848]: I1204 15:41:21.394111 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:41:21 crc kubenswrapper[4848]: E1204 15:41:21.394854 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:41:33 crc kubenswrapper[4848]: I1204 15:41:33.480290 4848 generic.go:334] "Generic (PLEG): container finished" podID="0c931889-7884-40ab-a9f4-8b90e49fa8e6" containerID="16adfe1c2595a3dea3f34e4665f00ffa325bbc3cf5d1fc04e3618212fd17d281" exitCode=0 Dec 04 15:41:33 crc kubenswrapper[4848]: I1204 15:41:33.480747 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/crc-debug-tq475" event={"ID":"0c931889-7884-40ab-a9f4-8b90e49fa8e6","Type":"ContainerDied","Data":"16adfe1c2595a3dea3f34e4665f00ffa325bbc3cf5d1fc04e3618212fd17d281"} Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.631917 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.676997 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9k646/crc-debug-tq475"] Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.686730 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9k646/crc-debug-tq475"] Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.759505 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0c931889-7884-40ab-a9f4-8b90e49fa8e6-host\") pod \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\" (UID: \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\") " Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.759605 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d85w\" (UniqueName: \"kubernetes.io/projected/0c931889-7884-40ab-a9f4-8b90e49fa8e6-kube-api-access-2d85w\") pod \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\" (UID: \"0c931889-7884-40ab-a9f4-8b90e49fa8e6\") " Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.759736 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c931889-7884-40ab-a9f4-8b90e49fa8e6-host" (OuterVolumeSpecName: "host") pod "0c931889-7884-40ab-a9f4-8b90e49fa8e6" (UID: "0c931889-7884-40ab-a9f4-8b90e49fa8e6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.760665 4848 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0c931889-7884-40ab-a9f4-8b90e49fa8e6-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.767500 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c931889-7884-40ab-a9f4-8b90e49fa8e6-kube-api-access-2d85w" (OuterVolumeSpecName: "kube-api-access-2d85w") pod "0c931889-7884-40ab-a9f4-8b90e49fa8e6" (UID: "0c931889-7884-40ab-a9f4-8b90e49fa8e6"). InnerVolumeSpecName "kube-api-access-2d85w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:41:34 crc kubenswrapper[4848]: I1204 15:41:34.863894 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d85w\" (UniqueName: \"kubernetes.io/projected/0c931889-7884-40ab-a9f4-8b90e49fa8e6-kube-api-access-2d85w\") on node \"crc\" DevicePath \"\"" Dec 04 15:41:35 crc kubenswrapper[4848]: I1204 15:41:35.501824 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7498c5d882be639d8c81ed6cfcdeca30dbd1740b3ce23119c4339318a6f6f20" Dec 04 15:41:35 crc kubenswrapper[4848]: I1204 15:41:35.501889 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-tq475" Dec 04 15:41:35 crc kubenswrapper[4848]: I1204 15:41:35.889214 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9k646/crc-debug-494dq"] Dec 04 15:41:35 crc kubenswrapper[4848]: E1204 15:41:35.889703 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c931889-7884-40ab-a9f4-8b90e49fa8e6" containerName="container-00" Dec 04 15:41:35 crc kubenswrapper[4848]: I1204 15:41:35.889715 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c931889-7884-40ab-a9f4-8b90e49fa8e6" containerName="container-00" Dec 04 15:41:35 crc kubenswrapper[4848]: I1204 15:41:35.890009 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c931889-7884-40ab-a9f4-8b90e49fa8e6" containerName="container-00" Dec 04 15:41:35 crc kubenswrapper[4848]: I1204 15:41:35.890859 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:35 crc kubenswrapper[4848]: I1204 15:41:35.990072 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/62d328d5-17d3-46dc-ad75-79294f84c9ca-host\") pod \"crc-debug-494dq\" (UID: \"62d328d5-17d3-46dc-ad75-79294f84c9ca\") " pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:35 crc kubenswrapper[4848]: I1204 15:41:35.990534 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fb9c\" (UniqueName: \"kubernetes.io/projected/62d328d5-17d3-46dc-ad75-79294f84c9ca-kube-api-access-4fb9c\") pod \"crc-debug-494dq\" (UID: \"62d328d5-17d3-46dc-ad75-79294f84c9ca\") " pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:36 crc kubenswrapper[4848]: I1204 15:41:36.092448 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fb9c\" (UniqueName: \"kubernetes.io/projected/62d328d5-17d3-46dc-ad75-79294f84c9ca-kube-api-access-4fb9c\") pod \"crc-debug-494dq\" (UID: \"62d328d5-17d3-46dc-ad75-79294f84c9ca\") " pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:36 crc kubenswrapper[4848]: I1204 15:41:36.092600 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/62d328d5-17d3-46dc-ad75-79294f84c9ca-host\") pod \"crc-debug-494dq\" (UID: \"62d328d5-17d3-46dc-ad75-79294f84c9ca\") " pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:36 crc kubenswrapper[4848]: I1204 15:41:36.092747 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/62d328d5-17d3-46dc-ad75-79294f84c9ca-host\") pod \"crc-debug-494dq\" (UID: \"62d328d5-17d3-46dc-ad75-79294f84c9ca\") " pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:36 crc kubenswrapper[4848]: I1204 15:41:36.116172 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fb9c\" (UniqueName: \"kubernetes.io/projected/62d328d5-17d3-46dc-ad75-79294f84c9ca-kube-api-access-4fb9c\") pod \"crc-debug-494dq\" (UID: \"62d328d5-17d3-46dc-ad75-79294f84c9ca\") " pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:36 crc kubenswrapper[4848]: I1204 15:41:36.210927 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:36 crc kubenswrapper[4848]: I1204 15:41:36.401359 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:41:36 crc kubenswrapper[4848]: E1204 15:41:36.401653 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:41:36 crc kubenswrapper[4848]: I1204 15:41:36.415993 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c931889-7884-40ab-a9f4-8b90e49fa8e6" path="/var/lib/kubelet/pods/0c931889-7884-40ab-a9f4-8b90e49fa8e6/volumes" Dec 04 15:41:36 crc kubenswrapper[4848]: I1204 15:41:36.514509 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/crc-debug-494dq" event={"ID":"62d328d5-17d3-46dc-ad75-79294f84c9ca","Type":"ContainerStarted","Data":"b25624d91c12450befb03315fd0c4c7f614fcc181bff905b1c34a4351ce25ba5"} Dec 04 15:41:37 crc kubenswrapper[4848]: I1204 15:41:37.526549 4848 generic.go:334] "Generic (PLEG): container finished" podID="62d328d5-17d3-46dc-ad75-79294f84c9ca" containerID="3982eba4321306a24b0b6c4d7ce674cfb14531ff094cf235878a118c24b3da90" exitCode=0 Dec 04 15:41:37 crc kubenswrapper[4848]: I1204 15:41:37.526649 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/crc-debug-494dq" event={"ID":"62d328d5-17d3-46dc-ad75-79294f84c9ca","Type":"ContainerDied","Data":"3982eba4321306a24b0b6c4d7ce674cfb14531ff094cf235878a118c24b3da90"} Dec 04 15:41:38 crc kubenswrapper[4848]: I1204 15:41:38.653415 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:38 crc kubenswrapper[4848]: I1204 15:41:38.750649 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fb9c\" (UniqueName: \"kubernetes.io/projected/62d328d5-17d3-46dc-ad75-79294f84c9ca-kube-api-access-4fb9c\") pod \"62d328d5-17d3-46dc-ad75-79294f84c9ca\" (UID: \"62d328d5-17d3-46dc-ad75-79294f84c9ca\") " Dec 04 15:41:38 crc kubenswrapper[4848]: I1204 15:41:38.750743 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/62d328d5-17d3-46dc-ad75-79294f84c9ca-host\") pod \"62d328d5-17d3-46dc-ad75-79294f84c9ca\" (UID: \"62d328d5-17d3-46dc-ad75-79294f84c9ca\") " Dec 04 15:41:38 crc kubenswrapper[4848]: I1204 15:41:38.750867 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62d328d5-17d3-46dc-ad75-79294f84c9ca-host" (OuterVolumeSpecName: "host") pod "62d328d5-17d3-46dc-ad75-79294f84c9ca" (UID: "62d328d5-17d3-46dc-ad75-79294f84c9ca"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:41:38 crc kubenswrapper[4848]: I1204 15:41:38.756035 4848 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/62d328d5-17d3-46dc-ad75-79294f84c9ca-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:41:38 crc kubenswrapper[4848]: I1204 15:41:38.759006 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d328d5-17d3-46dc-ad75-79294f84c9ca-kube-api-access-4fb9c" (OuterVolumeSpecName: "kube-api-access-4fb9c") pod "62d328d5-17d3-46dc-ad75-79294f84c9ca" (UID: "62d328d5-17d3-46dc-ad75-79294f84c9ca"). InnerVolumeSpecName "kube-api-access-4fb9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:41:38 crc kubenswrapper[4848]: I1204 15:41:38.858300 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fb9c\" (UniqueName: \"kubernetes.io/projected/62d328d5-17d3-46dc-ad75-79294f84c9ca-kube-api-access-4fb9c\") on node \"crc\" DevicePath \"\"" Dec 04 15:41:39 crc kubenswrapper[4848]: I1204 15:41:39.551430 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/crc-debug-494dq" event={"ID":"62d328d5-17d3-46dc-ad75-79294f84c9ca","Type":"ContainerDied","Data":"b25624d91c12450befb03315fd0c4c7f614fcc181bff905b1c34a4351ce25ba5"} Dec 04 15:41:39 crc kubenswrapper[4848]: I1204 15:41:39.551470 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b25624d91c12450befb03315fd0c4c7f614fcc181bff905b1c34a4351ce25ba5" Dec 04 15:41:39 crc kubenswrapper[4848]: I1204 15:41:39.551519 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-494dq" Dec 04 15:41:39 crc kubenswrapper[4848]: I1204 15:41:39.577666 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9k646/crc-debug-494dq"] Dec 04 15:41:39 crc kubenswrapper[4848]: I1204 15:41:39.587805 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9k646/crc-debug-494dq"] Dec 04 15:41:40 crc kubenswrapper[4848]: I1204 15:41:40.417542 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62d328d5-17d3-46dc-ad75-79294f84c9ca" path="/var/lib/kubelet/pods/62d328d5-17d3-46dc-ad75-79294f84c9ca/volumes" Dec 04 15:41:40 crc kubenswrapper[4848]: I1204 15:41:40.928061 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9k646/crc-debug-nxt8d"] Dec 04 15:41:40 crc kubenswrapper[4848]: E1204 15:41:40.928558 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d328d5-17d3-46dc-ad75-79294f84c9ca" containerName="container-00" Dec 04 15:41:40 crc kubenswrapper[4848]: I1204 15:41:40.928575 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d328d5-17d3-46dc-ad75-79294f84c9ca" containerName="container-00" Dec 04 15:41:40 crc kubenswrapper[4848]: I1204 15:41:40.928822 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d328d5-17d3-46dc-ad75-79294f84c9ca" containerName="container-00" Dec 04 15:41:40 crc kubenswrapper[4848]: I1204 15:41:40.929567 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:40 crc kubenswrapper[4848]: I1204 15:41:40.949918 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf6s5\" (UniqueName: \"kubernetes.io/projected/46756e62-6bf1-4f3a-b647-d1509a5aa1de-kube-api-access-sf6s5\") pod \"crc-debug-nxt8d\" (UID: \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\") " pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:40 crc kubenswrapper[4848]: I1204 15:41:40.950119 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46756e62-6bf1-4f3a-b647-d1509a5aa1de-host\") pod \"crc-debug-nxt8d\" (UID: \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\") " pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.051260 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46756e62-6bf1-4f3a-b647-d1509a5aa1de-host\") pod \"crc-debug-nxt8d\" (UID: \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\") " pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.051383 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46756e62-6bf1-4f3a-b647-d1509a5aa1de-host\") pod \"crc-debug-nxt8d\" (UID: \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\") " pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.051431 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf6s5\" (UniqueName: \"kubernetes.io/projected/46756e62-6bf1-4f3a-b647-d1509a5aa1de-kube-api-access-sf6s5\") pod \"crc-debug-nxt8d\" (UID: \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\") " pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.073847 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf6s5\" (UniqueName: \"kubernetes.io/projected/46756e62-6bf1-4f3a-b647-d1509a5aa1de-kube-api-access-sf6s5\") pod \"crc-debug-nxt8d\" (UID: \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\") " pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.257343 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:41 crc kubenswrapper[4848]: W1204 15:41:41.298761 4848 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46756e62_6bf1_4f3a_b647_d1509a5aa1de.slice/crio-e1e159a3646ba995b73ed14daa3d7ec3c3b2d4525563d44b0c3749642ee8588d WatchSource:0}: Error finding container e1e159a3646ba995b73ed14daa3d7ec3c3b2d4525563d44b0c3749642ee8588d: Status 404 returned error can't find the container with id e1e159a3646ba995b73ed14daa3d7ec3c3b2d4525563d44b0c3749642ee8588d Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.877825 4848 generic.go:334] "Generic (PLEG): container finished" podID="46756e62-6bf1-4f3a-b647-d1509a5aa1de" containerID="05f85a6295921e54988fe8610e28c4def2e0969931d7a6d0838320867d5a4f46" exitCode=0 Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.877938 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/crc-debug-nxt8d" event={"ID":"46756e62-6bf1-4f3a-b647-d1509a5aa1de","Type":"ContainerDied","Data":"05f85a6295921e54988fe8610e28c4def2e0969931d7a6d0838320867d5a4f46"} Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.878230 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/crc-debug-nxt8d" event={"ID":"46756e62-6bf1-4f3a-b647-d1509a5aa1de","Type":"ContainerStarted","Data":"e1e159a3646ba995b73ed14daa3d7ec3c3b2d4525563d44b0c3749642ee8588d"} Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.936414 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9k646/crc-debug-nxt8d"] Dec 04 15:41:41 crc kubenswrapper[4848]: I1204 15:41:41.948567 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9k646/crc-debug-nxt8d"] Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.005364 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.197411 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf6s5\" (UniqueName: \"kubernetes.io/projected/46756e62-6bf1-4f3a-b647-d1509a5aa1de-kube-api-access-sf6s5\") pod \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\" (UID: \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\") " Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.197551 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46756e62-6bf1-4f3a-b647-d1509a5aa1de-host\") pod \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\" (UID: \"46756e62-6bf1-4f3a-b647-d1509a5aa1de\") " Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.197705 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46756e62-6bf1-4f3a-b647-d1509a5aa1de-host" (OuterVolumeSpecName: "host") pod "46756e62-6bf1-4f3a-b647-d1509a5aa1de" (UID: "46756e62-6bf1-4f3a-b647-d1509a5aa1de"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.198282 4848 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46756e62-6bf1-4f3a-b647-d1509a5aa1de-host\") on node \"crc\" DevicePath \"\"" Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.203512 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46756e62-6bf1-4f3a-b647-d1509a5aa1de-kube-api-access-sf6s5" (OuterVolumeSpecName: "kube-api-access-sf6s5") pod "46756e62-6bf1-4f3a-b647-d1509a5aa1de" (UID: "46756e62-6bf1-4f3a-b647-d1509a5aa1de"). InnerVolumeSpecName "kube-api-access-sf6s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.300498 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf6s5\" (UniqueName: \"kubernetes.io/projected/46756e62-6bf1-4f3a-b647-d1509a5aa1de-kube-api-access-sf6s5\") on node \"crc\" DevicePath \"\"" Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.901149 4848 scope.go:117] "RemoveContainer" containerID="05f85a6295921e54988fe8610e28c4def2e0969931d7a6d0838320867d5a4f46" Dec 04 15:41:43 crc kubenswrapper[4848]: I1204 15:41:43.901189 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/crc-debug-nxt8d" Dec 04 15:41:44 crc kubenswrapper[4848]: I1204 15:41:44.406486 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46756e62-6bf1-4f3a-b647-d1509a5aa1de" path="/var/lib/kubelet/pods/46756e62-6bf1-4f3a-b647-d1509a5aa1de/volumes" Dec 04 15:41:50 crc kubenswrapper[4848]: I1204 15:41:50.394097 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:41:51 crc kubenswrapper[4848]: I1204 15:41:51.007057 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"78749c589f78ef3cb53b6338248052af0028bcb29d8a4feb7475ba80de9997d6"} Dec 04 15:42:12 crc kubenswrapper[4848]: I1204 15:42:12.752337 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc646c01-afe4-4b6b-ab39-d8c333cc3ed4/aodh-api/0.log" Dec 04 15:42:12 crc kubenswrapper[4848]: I1204 15:42:12.836807 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc646c01-afe4-4b6b-ab39-d8c333cc3ed4/aodh-evaluator/0.log" Dec 04 15:42:12 crc kubenswrapper[4848]: I1204 15:42:12.979533 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc646c01-afe4-4b6b-ab39-d8c333cc3ed4/aodh-notifier/0.log" Dec 04 15:42:13 crc kubenswrapper[4848]: I1204 15:42:13.007060 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_bc646c01-afe4-4b6b-ab39-d8c333cc3ed4/aodh-listener/0.log" Dec 04 15:42:13 crc kubenswrapper[4848]: I1204 15:42:13.043797 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-786894c88b-2c6rg_5e05a6e1-667b-44bd-a532-e91c2245e876/barbican-api/0.log" Dec 04 15:42:13 crc kubenswrapper[4848]: I1204 15:42:13.205102 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-786894c88b-2c6rg_5e05a6e1-667b-44bd-a532-e91c2245e876/barbican-api-log/0.log" Dec 04 15:42:13 crc kubenswrapper[4848]: I1204 15:42:13.479223 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58957fbb8b-n55cg_c232050c-845f-4a98-b79d-1788830cfd67/barbican-keystone-listener/0.log" Dec 04 15:42:13 crc kubenswrapper[4848]: I1204 15:42:13.657891 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-58957fbb8b-n55cg_c232050c-845f-4a98-b79d-1788830cfd67/barbican-keystone-listener-log/0.log" Dec 04 15:42:13 crc kubenswrapper[4848]: I1204 15:42:13.688644 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b699c955-dtjdb_051c8366-c751-423e-a7c8-4d69b6a5897d/barbican-worker/0.log" Dec 04 15:42:13 crc kubenswrapper[4848]: I1204 15:42:13.748811 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b699c955-dtjdb_051c8366-c751-423e-a7c8-4d69b6a5897d/barbican-worker-log/0.log" Dec 04 15:42:13 crc kubenswrapper[4848]: I1204 15:42:13.927890 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-kkmk5_17b7cb97-4979-468f-89be-d14529b93be8/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.074293 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40c4de59-893b-42be-b67d-f3f067b410ac/ceilometer-central-agent/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.195349 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40c4de59-893b-42be-b67d-f3f067b410ac/ceilometer-notification-agent/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.203029 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40c4de59-893b-42be-b67d-f3f067b410ac/sg-core/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.214227 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40c4de59-893b-42be-b67d-f3f067b410ac/proxy-httpd/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.403171 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e0893eec-f2cb-41b8-a301-005dbf326afe/cinder-api-log/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.487796 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e0893eec-f2cb-41b8-a301-005dbf326afe/cinder-api/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.659480 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c93bea85-4d93-4003-b146-70a07e1edd92/cinder-scheduler/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.674863 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c93bea85-4d93-4003-b146-70a07e1edd92/probe/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.781016 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-qzsx9_fbf57421-8b0a-42bd-a682-65afd34f39bb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:14 crc kubenswrapper[4848]: I1204 15:42:14.951448 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-p5mgt_8c603058-8c35-4c40-b700-2c358d0a5df7/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:15 crc kubenswrapper[4848]: I1204 15:42:15.008065 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-g7qqg_7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd/init/0.log" Dec 04 15:42:15 crc kubenswrapper[4848]: I1204 15:42:15.205477 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-g7qqg_7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd/init/0.log" Dec 04 15:42:15 crc kubenswrapper[4848]: I1204 15:42:15.311001 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-g7qqg_7c7f9cd2-1789-486a-ac35-4dfaefeb6bbd/dnsmasq-dns/0.log" Dec 04 15:42:15 crc kubenswrapper[4848]: I1204 15:42:15.320872 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-9tmwp_9d69a925-9c1f-4405-9300-e6c4c3cada23/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:15 crc kubenswrapper[4848]: I1204 15:42:15.544800 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d5008baa-3c8b-4048-a87e-4e6f3e5d2327/glance-httpd/0.log" Dec 04 15:42:15 crc kubenswrapper[4848]: I1204 15:42:15.596203 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d5008baa-3c8b-4048-a87e-4e6f3e5d2327/glance-log/0.log" Dec 04 15:42:15 crc kubenswrapper[4848]: I1204 15:42:15.744699 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0c33bbab-5535-4d34-8b87-90c7e1549281/glance-log/0.log" Dec 04 15:42:15 crc kubenswrapper[4848]: I1204 15:42:15.755589 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0c33bbab-5535-4d34-8b87-90c7e1549281/glance-httpd/0.log" Dec 04 15:42:16 crc kubenswrapper[4848]: I1204 15:42:16.378124 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-55957f9948-8gknv_01ecc04b-d6b6-4182-ba10-0b2d14b6bef5/heat-engine/0.log" Dec 04 15:42:16 crc kubenswrapper[4848]: I1204 15:42:16.684090 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-m7sgp_fe59edb1-8c86-40a3-b68a-7f597af859de/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:16 crc kubenswrapper[4848]: I1204 15:42:16.738250 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6c746cbd45-fc8bq_4ab14888-fa03-4bbd-a567-95e9078fbe3b/heat-api/0.log" Dec 04 15:42:16 crc kubenswrapper[4848]: I1204 15:42:16.992334 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-gbtwr_51fab24f-2ea8-430c-810b-a538832876de/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:17 crc kubenswrapper[4848]: I1204 15:42:17.098857 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-6b8f46df8d-vvlcd_cfffd13e-a23e-41bd-b830-0feab48494ec/heat-cfnapi/0.log" Dec 04 15:42:17 crc kubenswrapper[4848]: I1204 15:42:17.277546 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29414341-72l4q_9fe270e6-d360-4e3f-851e-3ab6b76b92a9/keystone-cron/0.log" Dec 04 15:42:17 crc kubenswrapper[4848]: I1204 15:42:17.514559 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1d88c8b4-d31b-45d7-8ae5-99b50d716ae0/kube-state-metrics/0.log" Dec 04 15:42:17 crc kubenswrapper[4848]: I1204 15:42:17.659112 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-6lph2_3146cd9a-c65d-4942-bb4d-ee0951403d87/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:17 crc kubenswrapper[4848]: I1204 15:42:17.761481 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-9zcjg_d0800b88-1124-4b08-a53c-ba88de0acc98/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:17 crc kubenswrapper[4848]: I1204 15:42:17.776983 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5b74ff8c74-nm59b_6a28b540-21e3-43eb-9b64-215d661d4721/keystone-api/0.log" Dec 04 15:42:18 crc kubenswrapper[4848]: I1204 15:42:18.026798 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_2ddeb253-877f-4e3e-8d1c-b51c1621bd15/mysqld-exporter/0.log" Dec 04 15:42:18 crc kubenswrapper[4848]: I1204 15:42:18.344243 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjbcr_b1231a87-3a32-4b52-a5b0-d546f15523e1/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:18 crc kubenswrapper[4848]: I1204 15:42:18.451929 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b6b4654c9-hg8j8_59773efa-83b9-4645-8187-fc0ba4c1d70c/neutron-httpd/0.log" Dec 04 15:42:18 crc kubenswrapper[4848]: I1204 15:42:18.476615 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b6b4654c9-hg8j8_59773efa-83b9-4645-8187-fc0ba4c1d70c/neutron-api/0.log" Dec 04 15:42:19 crc kubenswrapper[4848]: I1204 15:42:19.146791 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7ae44733-c46b-4ed7-9371-377d026bc216/nova-cell0-conductor-conductor/0.log" Dec 04 15:42:19 crc kubenswrapper[4848]: I1204 15:42:19.482523 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d9781260-81d8-4950-8388-00ae237a6f7f/nova-api-log/0.log" Dec 04 15:42:19 crc kubenswrapper[4848]: I1204 15:42:19.485748 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fd31580e-d25b-49af-a7c6-2eb44a77af46/nova-cell1-conductor-conductor/0.log" Dec 04 15:42:19 crc kubenswrapper[4848]: I1204 15:42:19.838639 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_1c49601a-f35b-41cb-bbb6-d7a4dd0542c0/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 15:42:19 crc kubenswrapper[4848]: I1204 15:42:19.840328 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-v9fxp_86fbd978-880e-415b-ad1e-e367a0c0079b/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:19 crc kubenswrapper[4848]: I1204 15:42:19.985147 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d9781260-81d8-4950-8388-00ae237a6f7f/nova-api-api/0.log" Dec 04 15:42:20 crc kubenswrapper[4848]: I1204 15:42:20.178492 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_42fe17d3-9143-40e0-ab66-074885ccd028/nova-metadata-log/0.log" Dec 04 15:42:20 crc kubenswrapper[4848]: I1204 15:42:20.442974 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2/mysql-bootstrap/0.log" Dec 04 15:42:20 crc kubenswrapper[4848]: I1204 15:42:20.491403 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7fa9a476-cd4b-4258-ba07-0ef2888fa4e3/nova-scheduler-scheduler/0.log" Dec 04 15:42:20 crc kubenswrapper[4848]: I1204 15:42:20.671094 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2/mysql-bootstrap/0.log" Dec 04 15:42:20 crc kubenswrapper[4848]: I1204 15:42:20.703211 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1317a7f1-1cd4-4b84-aa1f-0e0b7d3a76e2/galera/0.log" Dec 04 15:42:21 crc kubenswrapper[4848]: I1204 15:42:21.214487 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52/mysql-bootstrap/0.log" Dec 04 15:42:21 crc kubenswrapper[4848]: I1204 15:42:21.309667 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52/mysql-bootstrap/0.log" Dec 04 15:42:21 crc kubenswrapper[4848]: I1204 15:42:21.321125 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f4e5ee36-ea81-42cf-b0c4-44f2dcd4cd52/galera/0.log" Dec 04 15:42:21 crc kubenswrapper[4848]: I1204 15:42:21.511035 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ec3f1251-eb61-4b0e-b534-177c7d928404/openstackclient/0.log" Dec 04 15:42:21 crc kubenswrapper[4848]: I1204 15:42:21.715771 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-b9f5p_caf1c526-2dfe-4c81-9846-c122113ebaf1/openstack-network-exporter/0.log" Dec 04 15:42:21 crc kubenswrapper[4848]: I1204 15:42:21.894814 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cpcmr_0ef872ad-6863-45f9-ac22-86841dea2bbf/ovsdb-server-init/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.086425 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cpcmr_0ef872ad-6863-45f9-ac22-86841dea2bbf/ovsdb-server-init/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.135286 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cpcmr_0ef872ad-6863-45f9-ac22-86841dea2bbf/ovs-vswitchd/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.166443 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cpcmr_0ef872ad-6863-45f9-ac22-86841dea2bbf/ovsdb-server/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.344768 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-rg7xl_ef9a0101-72b3-4879-a93b-f58362d175ed/ovn-controller/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.651336 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-l9mmb_54ce5df0-f586-463e-ad53-6ddc68e8ce90/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.667957 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c7da8d50-9ec2-4b1d-b373-e65541b02306/openstack-network-exporter/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.865254 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_12b7748e-3169-49a0-a634-8a485f7690f4/openstack-network-exporter/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.874856 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c7da8d50-9ec2-4b1d-b373-e65541b02306/ovn-northd/0.log" Dec 04 15:42:22 crc kubenswrapper[4848]: I1204 15:42:22.914548 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_42fe17d3-9143-40e0-ab66-074885ccd028/nova-metadata-metadata/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.062018 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_12b7748e-3169-49a0-a634-8a485f7690f4/ovsdbserver-nb/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.178605 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_97bf1fd2-3a36-418d-826a-6dc60889b70c/openstack-network-exporter/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.202113 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_97bf1fd2-3a36-418d-826a-6dc60889b70c/ovsdbserver-sb/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.609151 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/init-config-reloader/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.646307 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8766cc68-d2824_7c8a40cb-aac2-4a08-96f3-79c97ad82007/placement-api/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.651795 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8766cc68-d2824_7c8a40cb-aac2-4a08-96f3-79c97ad82007/placement-log/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.833583 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/init-config-reloader/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.868567 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/config-reloader/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.901622 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/thanos-sidecar/0.log" Dec 04 15:42:23 crc kubenswrapper[4848]: I1204 15:42:23.949260 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_a9e4a10a-a31e-431d-99e2-d4205440fad6/prometheus/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.059599 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6c94605c-497c-47b2-b606-22f72e1bd0cc/setup-container/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.275395 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6c94605c-497c-47b2-b606-22f72e1bd0cc/setup-container/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.364403 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a33214fe-5596-499a-9563-0931c49821c3/setup-container/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.392161 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6c94605c-497c-47b2-b606-22f72e1bd0cc/rabbitmq/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.568227 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a33214fe-5596-499a-9563-0931c49821c3/setup-container/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.645800 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a33214fe-5596-499a-9563-0931c49821c3/rabbitmq/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.689416 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-sb6lf_2af1d16b-4257-4fcd-b6fa-dc37b1313e40/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.879892 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-kwkbw_25e07c34-72be-4c3a-8432-3fb295a46d57/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:24 crc kubenswrapper[4848]: I1204 15:42:24.922707 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-62zpz_66ec76cb-b6b1-4846-83d3-58cf60fd2d3d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:25 crc kubenswrapper[4848]: I1204 15:42:25.129425 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dfmnc_5b8a5833-af2e-4fef-9bfa-bbcec7064538/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:25 crc kubenswrapper[4848]: I1204 15:42:25.221629 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dxn92_7ae92c99-2f33-44d5-a478-9a685afd0f89/ssh-known-hosts-edpm-deployment/0.log" Dec 04 15:42:25 crc kubenswrapper[4848]: I1204 15:42:25.686920 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b65c5f4cf-q22f8_acf82280-f41a-460a-84ab-639c541670fc/proxy-server/0.log" Dec 04 15:42:25 crc kubenswrapper[4848]: I1204 15:42:25.797712 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gplgt_ff826431-be61-4dea-82e7-b6849d838e52/swift-ring-rebalance/0.log" Dec 04 15:42:25 crc kubenswrapper[4848]: I1204 15:42:25.885053 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-b65c5f4cf-q22f8_acf82280-f41a-460a-84ab-639c541670fc/proxy-httpd/0.log" Dec 04 15:42:25 crc kubenswrapper[4848]: I1204 15:42:25.921723 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/account-auditor/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.116421 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/account-reaper/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.202574 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/account-server/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.263507 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/account-replicator/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.276539 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/container-auditor/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.420251 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/container-server/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.430748 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/container-replicator/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.480048 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/container-updater/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.569703 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-auditor/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.641114 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-expirer/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.735750 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-server/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.795764 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-updater/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.813105 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/object-replicator/0.log" Dec 04 15:42:26 crc kubenswrapper[4848]: I1204 15:42:26.855585 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/rsync/0.log" Dec 04 15:42:27 crc kubenswrapper[4848]: I1204 15:42:27.017754 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0102871d-c950-46e7-af43-a2b3f1cbc39f/swift-recon-cron/0.log" Dec 04 15:42:27 crc kubenswrapper[4848]: I1204 15:42:27.109459 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-k5pzs_d5bb9916-94fb-4c6a-ad02-d2c85a1dafd1/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:27 crc kubenswrapper[4848]: I1204 15:42:27.404081 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-btmbg_0e112a8b-6965-4e32-bb39-4ff08519bdee/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:27 crc kubenswrapper[4848]: I1204 15:42:27.540238 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_3d34d919-565d-49be-8170-1bec9816c2c9/test-operator-logs-container/0.log" Dec 04 15:42:27 crc kubenswrapper[4848]: I1204 15:42:27.692035 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-bm7jt_bae5190f-858d-4fdc-8116-78911f113311/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 15:42:28 crc kubenswrapper[4848]: I1204 15:42:28.640922 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_9c0a7237-5ae8-46ef-9d7a-6aa204c6ed5a/tempest-tests-tempest-tests-runner/0.log" Dec 04 15:42:44 crc kubenswrapper[4848]: I1204 15:42:44.022441 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_add25985-bc0c-48ee-9f42-4404503638b5/memcached/0.log" Dec 04 15:42:57 crc kubenswrapper[4848]: I1204 15:42:57.672399 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-65bvj_4ad79d92-0097-4c46-930a-75489fad9915/kube-rbac-proxy/0.log" Dec 04 15:42:57 crc kubenswrapper[4848]: I1204 15:42:57.917338 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-9kj8j_038e3294-a0db-4c23-8ef2-ed6a7e617a3f/kube-rbac-proxy/0.log" Dec 04 15:42:57 crc kubenswrapper[4848]: I1204 15:42:57.935417 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-65bvj_4ad79d92-0097-4c46-930a-75489fad9915/manager/0.log" Dec 04 15:42:57 crc kubenswrapper[4848]: I1204 15:42:57.942617 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-9kj8j_038e3294-a0db-4c23-8ef2-ed6a7e617a3f/manager/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.124906 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-pv6z2_7a7135b8-89ed-4a8a-a09e-7e343bcd30ba/kube-rbac-proxy/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.182518 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-pv6z2_7a7135b8-89ed-4a8a-a09e-7e343bcd30ba/manager/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.339805 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/util/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.474102 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/pull/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.512683 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/util/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.528371 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/pull/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.708717 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/util/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.709633 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/pull/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.732429 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fb6485d729ddb9c562f56dd05b156706368cd25f32644ad76730a4cff8gbmph_ca659f4e-033e-45a9-8d3f-898c3fdf51cb/extract/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.899213 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-zljs9_7b00aa45-21c2-4be2-b702-8f191fc12b24/kube-rbac-proxy/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.974058 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4zcdk_1db7d065-38b0-4ef5-a468-660c4073f7f6/kube-rbac-proxy/0.log" Dec 04 15:42:58 crc kubenswrapper[4848]: I1204 15:42:58.977228 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-zljs9_7b00aa45-21c2-4be2-b702-8f191fc12b24/manager/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.164511 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4zcdk_1db7d065-38b0-4ef5-a468-660c4073f7f6/manager/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.205186 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pkb5l_a9c2feb0-4b79-4adc-b7a6-885a5a33c244/kube-rbac-proxy/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.218244 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pkb5l_a9c2feb0-4b79-4adc-b7a6-885a5a33c244/manager/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.357355 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-7g8w2_c22be20e-3619-45f0-944a-eff494ac7f1a/kube-rbac-proxy/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.556651 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-7g8w2_c22be20e-3619-45f0-944a-eff494ac7f1a/manager/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.564088 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-4hmhf_2fc8dada-7c7b-494c-ae6d-72c2737c6f66/kube-rbac-proxy/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.617147 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-4hmhf_2fc8dada-7c7b-494c-ae6d-72c2737c6f66/manager/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.745406 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kvfn2_de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a/kube-rbac-proxy/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.827720 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-kvfn2_de3bcc75-e7c5-4dc5-ac8d-8c14d21d275a/manager/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.845509 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-925th_2c849714-e889-446a-9b2c-6862cdbaf422/kube-rbac-proxy/0.log" Dec 04 15:42:59 crc kubenswrapper[4848]: I1204 15:42:59.961112 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-925th_2c849714-e889-446a-9b2c-6862cdbaf422/manager/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.018661 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-9sqtl_82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93/kube-rbac-proxy/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.098673 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-9sqtl_82b2e240-e4b4-4e4c-bfba-f00e3d8d7c93/manager/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.228224 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rr887_16e8aa39-31cb-4051-bdcf-797b4d5792c3/kube-rbac-proxy/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.251853 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-rr887_16e8aa39-31cb-4051-bdcf-797b4d5792c3/manager/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.421399 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-dvhkz_9e10ea48-b397-4f33-9da2-1a82a2987a9e/kube-rbac-proxy/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.476215 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-dvhkz_9e10ea48-b397-4f33-9da2-1a82a2987a9e/manager/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.513136 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-gdsqn_fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7/kube-rbac-proxy/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.687537 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd46t68x_a60c0a88-4168-4971-8bdd-638030697efb/kube-rbac-proxy/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.688128 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd46t68x_a60c0a88-4168-4971-8bdd-638030697efb/manager/0.log" Dec 04 15:43:00 crc kubenswrapper[4848]: I1204 15:43:00.689425 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-gdsqn_fe7a9851-7fb6-4b99-8cc3-61cdd471e5e7/manager/0.log" Dec 04 15:43:01 crc kubenswrapper[4848]: I1204 15:43:01.117371 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-799774d886-rj5dt_6641bd6c-020c-469a-8439-4953ffa85ee9/operator/0.log" Dec 04 15:43:01 crc kubenswrapper[4848]: I1204 15:43:01.203957 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-st8tz_409845c7-53dc-40ea-902b-b36b35b7d686/registry-server/0.log" Dec 04 15:43:01 crc kubenswrapper[4848]: I1204 15:43:01.415937 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-8hd8b_14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6/kube-rbac-proxy/0.log" Dec 04 15:43:01 crc kubenswrapper[4848]: I1204 15:43:01.568349 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-8hd8b_14b4ecc0-423c-46b0-928c-6d7fcfc2d8f6/manager/0.log" Dec 04 15:43:01 crc kubenswrapper[4848]: I1204 15:43:01.682818 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c4j2k_f15b3026-2d34-491c-a37f-c0053e3171ee/kube-rbac-proxy/0.log" Dec 04 15:43:01 crc kubenswrapper[4848]: I1204 15:43:01.852793 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c4j2k_f15b3026-2d34-491c-a37f-c0053e3171ee/manager/0.log" Dec 04 15:43:01 crc kubenswrapper[4848]: I1204 15:43:01.931500 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5l2qh_0488ea60-3643-4e40-94cd-555ac6c93043/operator/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.078470 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-kdftj_82a80725-8983-447c-8f98-d4b298ae8e7e/kube-rbac-proxy/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.165447 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-kdftj_82a80725-8983-447c-8f98-d4b298ae8e7e/manager/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.181218 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f78686648-gktm6_2c5d62fd-6700-4804-ba6d-4d9884b9ea02/kube-rbac-proxy/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.357286 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-576c48d787-gx8wm_12302e73-00be-4ffb-9cda-3ac6126e2f0c/manager/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.461754 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z2wqt_12439af0-d3a3-47d8-9626-b7e357af2100/kube-rbac-proxy/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.520973 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z2wqt_12439af0-d3a3-47d8-9626-b7e357af2100/manager/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.654127 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6f78686648-gktm6_2c5d62fd-6700-4804-ba6d-4d9884b9ea02/manager/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.666980 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-sr98b_5382b606-8cea-4b64-98ac-950f99c2a03d/kube-rbac-proxy/0.log" Dec 04 15:43:02 crc kubenswrapper[4848]: I1204 15:43:02.726235 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-sr98b_5382b606-8cea-4b64-98ac-950f99c2a03d/manager/0.log" Dec 04 15:43:20 crc kubenswrapper[4848]: I1204 15:43:20.196222 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qx7kz_41654867-fa7f-43c7-b672-c3eb6500b16c/control-plane-machine-set-operator/0.log" Dec 04 15:43:20 crc kubenswrapper[4848]: I1204 15:43:20.352082 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tk6sb_e8f7ac81-45f7-4446-a6df-78123ca7226b/machine-api-operator/0.log" Dec 04 15:43:20 crc kubenswrapper[4848]: I1204 15:43:20.372619 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tk6sb_e8f7ac81-45f7-4446-a6df-78123ca7226b/kube-rbac-proxy/0.log" Dec 04 15:43:32 crc kubenswrapper[4848]: I1204 15:43:32.001936 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-wcntc_327abad7-b085-4c01-8ecd-e287bda9f6ee/cert-manager-controller/0.log" Dec 04 15:43:32 crc kubenswrapper[4848]: I1204 15:43:32.449387 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-r78j7_d708a779-4728-4230-ba68-02eb26b4dc72/cert-manager-cainjector/0.log" Dec 04 15:43:32 crc kubenswrapper[4848]: I1204 15:43:32.496848 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-76vk7_7e550e0e-d6e6-4786-ae2b-46a4f1534877/cert-manager-webhook/0.log" Dec 04 15:43:45 crc kubenswrapper[4848]: I1204 15:43:45.110769 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-qj7vf_357ea823-946e-407d-8940-b430c72e324e/nmstate-console-plugin/0.log" Dec 04 15:43:45 crc kubenswrapper[4848]: I1204 15:43:45.278024 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-rggz7_973c2430-1dce-45e2-b08f-f87439b80cce/nmstate-handler/0.log" Dec 04 15:43:45 crc kubenswrapper[4848]: I1204 15:43:45.352536 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-4nd5h_afaef749-1755-4749-bb8f-64b3fcb3f768/nmstate-metrics/0.log" Dec 04 15:43:45 crc kubenswrapper[4848]: I1204 15:43:45.364699 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-4nd5h_afaef749-1755-4749-bb8f-64b3fcb3f768/kube-rbac-proxy/0.log" Dec 04 15:43:45 crc kubenswrapper[4848]: I1204 15:43:45.548163 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-69mx9_33caa45e-2061-41cc-8dca-94a382ffa50a/nmstate-operator/0.log" Dec 04 15:43:45 crc kubenswrapper[4848]: I1204 15:43:45.604488 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-sjwbx_ba09ba52-535f-41bc-adb2-210ca566a77a/nmstate-webhook/0.log" Dec 04 15:43:59 crc kubenswrapper[4848]: I1204 15:43:59.583493 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7ddb79586-x6vmg_461d1ece-26ff-40ed-a92c-f21300adeec7/kube-rbac-proxy/0.log" Dec 04 15:43:59 crc kubenswrapper[4848]: I1204 15:43:59.676520 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7ddb79586-x6vmg_461d1ece-26ff-40ed-a92c-f21300adeec7/manager/0.log" Dec 04 15:44:13 crc kubenswrapper[4848]: I1204 15:44:13.543289 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-lgmcj_5d3f5a36-9e68-4f3d-94b8-8cb59192211b/cluster-logging-operator/0.log" Dec 04 15:44:13 crc kubenswrapper[4848]: I1204 15:44:13.723686 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-wgpm5_13d80a8f-6a05-4fdb-99bd-6aaf5f563f87/collector/0.log" Dec 04 15:44:13 crc kubenswrapper[4848]: I1204 15:44:13.772373 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_b53d48d7-e9d7-442d-b2b8-044a7eb9e692/loki-compactor/0.log" Dec 04 15:44:13 crc kubenswrapper[4848]: I1204 15:44:13.930253 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-mkvbk_cf9c3660-b2e2-4fac-a328-832144acf92c/loki-distributor/0.log" Dec 04 15:44:13 crc kubenswrapper[4848]: I1204 15:44:13.960302 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-7db4f94957-p9qcz_3457940c-ae43-4e49-b616-a450d524ffd9/gateway/0.log" Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.027821 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-7db4f94957-p9qcz_3457940c-ae43-4e49-b616-a450d524ffd9/opa/0.log" Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.135720 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-7db4f94957-qqjgk_bf4fb005-0de1-45b7-95a1-0fc7783e3a72/gateway/0.log" Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.196598 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-7db4f94957-qqjgk_bf4fb005-0de1-45b7-95a1-0fc7783e3a72/opa/0.log" Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.314698 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.314752 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.319902 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_25f82b9f-ffd8-47b2-9abe-593234820cb5/loki-index-gateway/0.log" Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.447353 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_19d88d7c-8641-499e-8fc1-8864af7feb3c/loki-ingester/0.log" Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.511285 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-g78k5_6d38276f-2908-496a-aa46-175984204c26/loki-querier/0.log" Dec 04 15:44:14 crc kubenswrapper[4848]: I1204 15:44:14.614570 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-htkln_d659950d-e89c-45fa-90c5-4526593ed819/loki-query-frontend/0.log" Dec 04 15:44:27 crc kubenswrapper[4848]: I1204 15:44:27.716644 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-h8bgt_e8c600d3-9a0f-4cab-9581-e75fa003c19d/kube-rbac-proxy/0.log" Dec 04 15:44:27 crc kubenswrapper[4848]: I1204 15:44:27.807380 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-h8bgt_e8c600d3-9a0f-4cab-9581-e75fa003c19d/controller/0.log" Dec 04 15:44:27 crc kubenswrapper[4848]: I1204 15:44:27.942421 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-frr-files/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.058268 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-frr-files/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.090582 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-reloader/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.126771 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-metrics/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.128094 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-reloader/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.302402 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-frr-files/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.347583 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-metrics/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.357327 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-reloader/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.361516 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-metrics/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.563592 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-reloader/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.582284 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-metrics/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.584482 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/cp-frr-files/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.586446 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/controller/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.746733 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/frr-metrics/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.764568 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/kube-rbac-proxy-frr/0.log" Dec 04 15:44:28 crc kubenswrapper[4848]: I1204 15:44:28.803824 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/kube-rbac-proxy/0.log" Dec 04 15:44:29 crc kubenswrapper[4848]: I1204 15:44:29.014165 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/reloader/0.log" Dec 04 15:44:29 crc kubenswrapper[4848]: I1204 15:44:29.057543 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-gpdjr_054c1d24-8144-490d-b917-391860e5704d/frr-k8s-webhook-server/0.log" Dec 04 15:44:29 crc kubenswrapper[4848]: I1204 15:44:29.274192 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-657d4c7649-z2tns_fe476aa8-681e-4dfb-9e65-58910c24c9cc/manager/0.log" Dec 04 15:44:29 crc kubenswrapper[4848]: I1204 15:44:29.504387 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c56865d68-9mdfg_ab8ac511-522e-4116-9c7b-ab61a4a6e018/webhook-server/0.log" Dec 04 15:44:29 crc kubenswrapper[4848]: I1204 15:44:29.613852 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pxcmt_e1998e0d-8a5c-4a7a-8008-598967ed2da0/kube-rbac-proxy/0.log" Dec 04 15:44:30 crc kubenswrapper[4848]: I1204 15:44:30.346762 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-pxcmt_e1998e0d-8a5c-4a7a-8008-598967ed2da0/speaker/0.log" Dec 04 15:44:30 crc kubenswrapper[4848]: I1204 15:44:30.733194 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j45lv_a5455d26-d1da-4e85-a556-a14cba35f8a9/frr/0.log" Dec 04 15:44:42 crc kubenswrapper[4848]: I1204 15:44:42.335800 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/util/0.log" Dec 04 15:44:42 crc kubenswrapper[4848]: I1204 15:44:42.516827 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/util/0.log" Dec 04 15:44:42 crc kubenswrapper[4848]: I1204 15:44:42.554076 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/pull/0.log" Dec 04 15:44:42 crc kubenswrapper[4848]: I1204 15:44:42.583020 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/pull/0.log" Dec 04 15:44:42 crc kubenswrapper[4848]: I1204 15:44:42.746974 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/util/0.log" Dec 04 15:44:42 crc kubenswrapper[4848]: I1204 15:44:42.789844 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/pull/0.log" Dec 04 15:44:42 crc kubenswrapper[4848]: I1204 15:44:42.815025 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8wd82k_2748bd61-6b69-4c1f-95dd-f8465b1b542b/extract/0.log" Dec 04 15:44:42 crc kubenswrapper[4848]: I1204 15:44:42.933843 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/util/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.093043 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/util/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.129544 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/pull/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.141054 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/pull/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.270183 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/util/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.334852 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/pull/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.364378 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkjrlr_2741f83b-175a-409b-b36e-5ba53521b9b4/extract/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.436524 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/util/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.643057 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/util/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.653427 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/pull/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.653809 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/pull/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.812836 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/pull/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.816412 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/util/0.log" Dec 04 15:44:43 crc kubenswrapper[4848]: I1204 15:44:43.840293 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210z9jbl_f76a919c-d711-434c-847f-20a25cecfd57/extract/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.153007 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/util/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.314543 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.314897 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.327602 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/pull/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.344442 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/pull/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.354308 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/util/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.562989 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/pull/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.585639 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/extract/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.602141 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fxrkxk_670575a5-b7cb-43bb-bfa9-2cf7666f3449/util/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.728503 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/util/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.934066 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/pull/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.939473 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/util/0.log" Dec 04 15:44:44 crc kubenswrapper[4848]: I1204 15:44:44.966287 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/pull/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.136223 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/util/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.148115 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/pull/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.170600 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pt5dn_9fa52fcb-11a4-4729-b5b8-4ce27d912310/extract/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.326726 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-utilities/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.499490 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-utilities/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.519775 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-content/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.521144 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-content/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.694257 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-content/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.722434 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/extract-utilities/0.log" Dec 04 15:44:45 crc kubenswrapper[4848]: I1204 15:44:45.974176 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-utilities/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.129222 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lq8mk_1085de76-81b1-4779-97aa-cbffc93aed49/registry-server/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.203856 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-utilities/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.244730 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-content/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.262577 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-content/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.388032 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-content/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.388475 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/extract-utilities/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.591596 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-nq9tw_cf33e30f-cf18-4cdd-8b35-84f835998ffd/marketplace-operator/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.639038 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-utilities/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.865077 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-utilities/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.883833 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-content/0.log" Dec 04 15:44:46 crc kubenswrapper[4848]: I1204 15:44:46.888116 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-content/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.095187 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-utilities/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.142271 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/extract-content/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.311408 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-utilities/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.496169 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-779vs_4cf50040-7cda-4267-afff-be6e9d288f7a/registry-server/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.554982 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-content/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.581994 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-content/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.659354 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7plv_16ba46bb-a73e-41e9-bff8-88b2b43e7a23/registry-server/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.674711 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-utilities/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.726818 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-content/0.log" Dec 04 15:44:47 crc kubenswrapper[4848]: I1204 15:44:47.746898 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/extract-utilities/0.log" Dec 04 15:44:48 crc kubenswrapper[4848]: I1204 15:44:48.038396 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zfpmh_b179aa9a-174a-42ba-a529-9f2aab63c1e9/registry-server/0.log" Dec 04 15:44:59 crc kubenswrapper[4848]: I1204 15:44:59.314982 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-mwx95_41de33f3-4fb3-48fa-988b-cb6f059cab9e/prometheus-operator/0.log" Dec 04 15:44:59 crc kubenswrapper[4848]: I1204 15:44:59.435584 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6fbbf78b67-7p99w_be2f3162-5720-4171-82c9-4df4a4244ec8/prometheus-operator-admission-webhook/0.log" Dec 04 15:44:59 crc kubenswrapper[4848]: I1204 15:44:59.484758 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6fbbf78b67-h9pwz_202cc8e2-1429-410d-8c7f-36062221f810/prometheus-operator-admission-webhook/0.log" Dec 04 15:44:59 crc kubenswrapper[4848]: I1204 15:44:59.649419 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-fl7j8_7532e1ac-73b7-4cea-ab3f-eac0c9dc27eb/operator/0.log" Dec 04 15:44:59 crc kubenswrapper[4848]: I1204 15:44:59.663429 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-49kl7_8efe0158-56d4-4526-9683-263fb1ce6f23/observability-ui-dashboards/0.log" Dec 04 15:44:59 crc kubenswrapper[4848]: I1204 15:44:59.794296 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-tn52h_d6694eaf-b146-407c-95b2-c52ccc355a77/perses-operator/0.log" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.227700 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs"] Dec 04 15:45:00 crc kubenswrapper[4848]: E1204 15:45:00.228792 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46756e62-6bf1-4f3a-b647-d1509a5aa1de" containerName="container-00" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.228913 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="46756e62-6bf1-4f3a-b647-d1509a5aa1de" containerName="container-00" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.229616 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="46756e62-6bf1-4f3a-b647-d1509a5aa1de" containerName="container-00" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.231472 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.234632 4848 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.235431 4848 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.281675 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/723926cc-7e02-499f-ade2-f46edd172e41-secret-volume\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.282106 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/723926cc-7e02-499f-ade2-f46edd172e41-config-volume\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.282336 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfmsv\" (UniqueName: \"kubernetes.io/projected/723926cc-7e02-499f-ade2-f46edd172e41-kube-api-access-pfmsv\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.373632 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs"] Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.385031 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/723926cc-7e02-499f-ade2-f46edd172e41-config-volume\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.385114 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfmsv\" (UniqueName: \"kubernetes.io/projected/723926cc-7e02-499f-ade2-f46edd172e41-kube-api-access-pfmsv\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.385254 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/723926cc-7e02-499f-ade2-f46edd172e41-secret-volume\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.386922 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/723926cc-7e02-499f-ade2-f46edd172e41-config-volume\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.393892 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/723926cc-7e02-499f-ade2-f46edd172e41-secret-volume\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.405977 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfmsv\" (UniqueName: \"kubernetes.io/projected/723926cc-7e02-499f-ade2-f46edd172e41-kube-api-access-pfmsv\") pod \"collect-profiles-29414385-k2fvs\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:00 crc kubenswrapper[4848]: I1204 15:45:00.553690 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:01 crc kubenswrapper[4848]: I1204 15:45:01.141637 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs"] Dec 04 15:45:02 crc kubenswrapper[4848]: I1204 15:45:02.087840 4848 generic.go:334] "Generic (PLEG): container finished" podID="723926cc-7e02-499f-ade2-f46edd172e41" containerID="d4b676dfcb69cb1c123facc4b8a6c03c54465a90800fc2795b58f81e2acaf928" exitCode=0 Dec 04 15:45:02 crc kubenswrapper[4848]: I1204 15:45:02.087985 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" event={"ID":"723926cc-7e02-499f-ade2-f46edd172e41","Type":"ContainerDied","Data":"d4b676dfcb69cb1c123facc4b8a6c03c54465a90800fc2795b58f81e2acaf928"} Dec 04 15:45:02 crc kubenswrapper[4848]: I1204 15:45:02.088242 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" event={"ID":"723926cc-7e02-499f-ade2-f46edd172e41","Type":"ContainerStarted","Data":"b8e59c5ce9b3a5e4d6ca3df63d192eb45ea6ce39818e5d86c20e10eb7115ba16"} Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.617274 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.680049 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/723926cc-7e02-499f-ade2-f46edd172e41-config-volume\") pod \"723926cc-7e02-499f-ade2-f46edd172e41\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.680555 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfmsv\" (UniqueName: \"kubernetes.io/projected/723926cc-7e02-499f-ade2-f46edd172e41-kube-api-access-pfmsv\") pod \"723926cc-7e02-499f-ade2-f46edd172e41\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.680768 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/723926cc-7e02-499f-ade2-f46edd172e41-secret-volume\") pod \"723926cc-7e02-499f-ade2-f46edd172e41\" (UID: \"723926cc-7e02-499f-ade2-f46edd172e41\") " Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.682963 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/723926cc-7e02-499f-ade2-f46edd172e41-config-volume" (OuterVolumeSpecName: "config-volume") pod "723926cc-7e02-499f-ade2-f46edd172e41" (UID: "723926cc-7e02-499f-ade2-f46edd172e41"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.691646 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/723926cc-7e02-499f-ade2-f46edd172e41-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "723926cc-7e02-499f-ade2-f46edd172e41" (UID: "723926cc-7e02-499f-ade2-f46edd172e41"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.691841 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/723926cc-7e02-499f-ade2-f46edd172e41-kube-api-access-pfmsv" (OuterVolumeSpecName: "kube-api-access-pfmsv") pod "723926cc-7e02-499f-ade2-f46edd172e41" (UID: "723926cc-7e02-499f-ade2-f46edd172e41"). InnerVolumeSpecName "kube-api-access-pfmsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.784277 4848 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/723926cc-7e02-499f-ade2-f46edd172e41-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.784319 4848 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/723926cc-7e02-499f-ade2-f46edd172e41-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 15:45:03 crc kubenswrapper[4848]: I1204 15:45:03.784332 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfmsv\" (UniqueName: \"kubernetes.io/projected/723926cc-7e02-499f-ade2-f46edd172e41-kube-api-access-pfmsv\") on node \"crc\" DevicePath \"\"" Dec 04 15:45:04 crc kubenswrapper[4848]: I1204 15:45:04.111674 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" event={"ID":"723926cc-7e02-499f-ade2-f46edd172e41","Type":"ContainerDied","Data":"b8e59c5ce9b3a5e4d6ca3df63d192eb45ea6ce39818e5d86c20e10eb7115ba16"} Dec 04 15:45:04 crc kubenswrapper[4848]: I1204 15:45:04.111771 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414385-k2fvs" Dec 04 15:45:04 crc kubenswrapper[4848]: I1204 15:45:04.112077 4848 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8e59c5ce9b3a5e4d6ca3df63d192eb45ea6ce39818e5d86c20e10eb7115ba16" Dec 04 15:45:04 crc kubenswrapper[4848]: I1204 15:45:04.713830 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr"] Dec 04 15:45:04 crc kubenswrapper[4848]: I1204 15:45:04.727632 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414340-5bpkr"] Dec 04 15:45:06 crc kubenswrapper[4848]: I1204 15:45:06.406810 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="663026c2-88c0-4a7f-b92f-392fb9d7afe7" path="/var/lib/kubelet/pods/663026c2-88c0-4a7f-b92f-392fb9d7afe7/volumes" Dec 04 15:45:12 crc kubenswrapper[4848]: I1204 15:45:12.025396 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7ddb79586-x6vmg_461d1ece-26ff-40ed-a92c-f21300adeec7/kube-rbac-proxy/0.log" Dec 04 15:45:12 crc kubenswrapper[4848]: I1204 15:45:12.030344 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-7ddb79586-x6vmg_461d1ece-26ff-40ed-a92c-f21300adeec7/manager/0.log" Dec 04 15:45:14 crc kubenswrapper[4848]: I1204 15:45:14.313916 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:45:14 crc kubenswrapper[4848]: I1204 15:45:14.314483 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:45:14 crc kubenswrapper[4848]: I1204 15:45:14.314531 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:45:14 crc kubenswrapper[4848]: I1204 15:45:14.316131 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"78749c589f78ef3cb53b6338248052af0028bcb29d8a4feb7475ba80de9997d6"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:45:14 crc kubenswrapper[4848]: I1204 15:45:14.316294 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://78749c589f78ef3cb53b6338248052af0028bcb29d8a4feb7475ba80de9997d6" gracePeriod=600 Dec 04 15:45:15 crc kubenswrapper[4848]: I1204 15:45:15.231856 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="78749c589f78ef3cb53b6338248052af0028bcb29d8a4feb7475ba80de9997d6" exitCode=0 Dec 04 15:45:15 crc kubenswrapper[4848]: I1204 15:45:15.231933 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"78749c589f78ef3cb53b6338248052af0028bcb29d8a4feb7475ba80de9997d6"} Dec 04 15:45:15 crc kubenswrapper[4848]: I1204 15:45:15.232516 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerStarted","Data":"b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d"} Dec 04 15:45:15 crc kubenswrapper[4848]: I1204 15:45:15.233040 4848 scope.go:117] "RemoveContainer" containerID="3151d77e83ad576f57546e74cd0c9cd9c8ad6e4f611782eff81b802fa4d1f760" Dec 04 15:45:18 crc kubenswrapper[4848]: E1204 15:45:18.045968 4848 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.5:42914->38.102.83.5:35999: write tcp 38.102.83.5:42914->38.102.83.5:35999: write: broken pipe Dec 04 15:45:26 crc kubenswrapper[4848]: E1204 15:45:26.430594 4848 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.5:55894->38.102.83.5:35999: write tcp 38.102.83.5:55894->38.102.83.5:35999: write: broken pipe Dec 04 15:45:34 crc kubenswrapper[4848]: E1204 15:45:34.849150 4848 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.5:35932->38.102.83.5:35999: write tcp 38.102.83.5:35932->38.102.83.5:35999: write: connection reset by peer Dec 04 15:45:42 crc kubenswrapper[4848]: I1204 15:45:42.870918 4848 scope.go:117] "RemoveContainer" containerID="87f332dca13d7bad2e654f1889b4af5b87fb7ef95904323e2c9513b42313887f" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.056549 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rm88h"] Dec 04 15:46:14 crc kubenswrapper[4848]: E1204 15:46:14.061004 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="723926cc-7e02-499f-ade2-f46edd172e41" containerName="collect-profiles" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.061053 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="723926cc-7e02-499f-ade2-f46edd172e41" containerName="collect-profiles" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.061721 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="723926cc-7e02-499f-ade2-f46edd172e41" containerName="collect-profiles" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.096847 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.155867 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-catalog-content\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.156140 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-utilities\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.156338 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhnxh\" (UniqueName: \"kubernetes.io/projected/0ea9a581-81d0-4255-9781-49ec01b78311-kube-api-access-jhnxh\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.211536 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rm88h"] Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.259049 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-utilities\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.259151 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhnxh\" (UniqueName: \"kubernetes.io/projected/0ea9a581-81d0-4255-9781-49ec01b78311-kube-api-access-jhnxh\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.259313 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-catalog-content\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.260103 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-catalog-content\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.260570 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-utilities\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.309852 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhnxh\" (UniqueName: \"kubernetes.io/projected/0ea9a581-81d0-4255-9781-49ec01b78311-kube-api-access-jhnxh\") pod \"redhat-marketplace-rm88h\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.430716 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.937927 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rm88h"] Dec 04 15:46:14 crc kubenswrapper[4848]: I1204 15:46:14.951249 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rm88h" event={"ID":"0ea9a581-81d0-4255-9781-49ec01b78311","Type":"ContainerStarted","Data":"80bd4268878a35d056fab6600789c645917917ec8ddebe16f2f25373598aeed8"} Dec 04 15:46:17 crc kubenswrapper[4848]: I1204 15:46:17.030296 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rm88h" event={"ID":"0ea9a581-81d0-4255-9781-49ec01b78311","Type":"ContainerStarted","Data":"fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a"} Dec 04 15:46:18 crc kubenswrapper[4848]: I1204 15:46:18.045648 4848 generic.go:334] "Generic (PLEG): container finished" podID="0ea9a581-81d0-4255-9781-49ec01b78311" containerID="fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a" exitCode=0 Dec 04 15:46:18 crc kubenswrapper[4848]: I1204 15:46:18.045930 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rm88h" event={"ID":"0ea9a581-81d0-4255-9781-49ec01b78311","Type":"ContainerDied","Data":"fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a"} Dec 04 15:46:18 crc kubenswrapper[4848]: I1204 15:46:18.050759 4848 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 15:46:19 crc kubenswrapper[4848]: I1204 15:46:19.058567 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rm88h" event={"ID":"0ea9a581-81d0-4255-9781-49ec01b78311","Type":"ContainerStarted","Data":"48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1"} Dec 04 15:46:19 crc kubenswrapper[4848]: E1204 15:46:19.429340 4848 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ea9a581_81d0_4255_9781_49ec01b78311.slice/crio-48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1.scope\": RecentStats: unable to find data in memory cache]" Dec 04 15:46:20 crc kubenswrapper[4848]: I1204 15:46:20.071152 4848 generic.go:334] "Generic (PLEG): container finished" podID="0ea9a581-81d0-4255-9781-49ec01b78311" containerID="48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1" exitCode=0 Dec 04 15:46:20 crc kubenswrapper[4848]: I1204 15:46:20.071193 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rm88h" event={"ID":"0ea9a581-81d0-4255-9781-49ec01b78311","Type":"ContainerDied","Data":"48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1"} Dec 04 15:46:22 crc kubenswrapper[4848]: I1204 15:46:22.095278 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rm88h" event={"ID":"0ea9a581-81d0-4255-9781-49ec01b78311","Type":"ContainerStarted","Data":"0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1"} Dec 04 15:46:22 crc kubenswrapper[4848]: I1204 15:46:22.116571 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rm88h" podStartSLOduration=4.853924255 podStartE2EDuration="8.116551276s" podCreationTimestamp="2025-12-04 15:46:14 +0000 UTC" firstStartedPulling="2025-12-04 15:46:18.048606618 +0000 UTC m=+7081.991103146" lastFinishedPulling="2025-12-04 15:46:21.311233649 +0000 UTC m=+7085.253730167" observedRunningTime="2025-12-04 15:46:22.110618341 +0000 UTC m=+7086.053114869" watchObservedRunningTime="2025-12-04 15:46:22.116551276 +0000 UTC m=+7086.059047804" Dec 04 15:46:24 crc kubenswrapper[4848]: I1204 15:46:24.432182 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:24 crc kubenswrapper[4848]: I1204 15:46:24.432574 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:24 crc kubenswrapper[4848]: I1204 15:46:24.491725 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:34 crc kubenswrapper[4848]: I1204 15:46:34.496912 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:34 crc kubenswrapper[4848]: I1204 15:46:34.549499 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rm88h"] Dec 04 15:46:35 crc kubenswrapper[4848]: I1204 15:46:35.253361 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rm88h" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" containerName="registry-server" containerID="cri-o://0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1" gracePeriod=2 Dec 04 15:46:35 crc kubenswrapper[4848]: I1204 15:46:35.796502 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:35 crc kubenswrapper[4848]: I1204 15:46:35.919206 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhnxh\" (UniqueName: \"kubernetes.io/projected/0ea9a581-81d0-4255-9781-49ec01b78311-kube-api-access-jhnxh\") pod \"0ea9a581-81d0-4255-9781-49ec01b78311\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " Dec 04 15:46:35 crc kubenswrapper[4848]: I1204 15:46:35.919410 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-utilities\") pod \"0ea9a581-81d0-4255-9781-49ec01b78311\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " Dec 04 15:46:35 crc kubenswrapper[4848]: I1204 15:46:35.919595 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-catalog-content\") pod \"0ea9a581-81d0-4255-9781-49ec01b78311\" (UID: \"0ea9a581-81d0-4255-9781-49ec01b78311\") " Dec 04 15:46:35 crc kubenswrapper[4848]: I1204 15:46:35.920191 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-utilities" (OuterVolumeSpecName: "utilities") pod "0ea9a581-81d0-4255-9781-49ec01b78311" (UID: "0ea9a581-81d0-4255-9781-49ec01b78311"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:46:35 crc kubenswrapper[4848]: I1204 15:46:35.927148 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea9a581-81d0-4255-9781-49ec01b78311-kube-api-access-jhnxh" (OuterVolumeSpecName: "kube-api-access-jhnxh") pod "0ea9a581-81d0-4255-9781-49ec01b78311" (UID: "0ea9a581-81d0-4255-9781-49ec01b78311"). InnerVolumeSpecName "kube-api-access-jhnxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:46:35 crc kubenswrapper[4848]: I1204 15:46:35.937040 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ea9a581-81d0-4255-9781-49ec01b78311" (UID: "0ea9a581-81d0-4255-9781-49ec01b78311"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.022156 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.022184 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea9a581-81d0-4255-9781-49ec01b78311-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.022212 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhnxh\" (UniqueName: \"kubernetes.io/projected/0ea9a581-81d0-4255-9781-49ec01b78311-kube-api-access-jhnxh\") on node \"crc\" DevicePath \"\"" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.265010 4848 generic.go:334] "Generic (PLEG): container finished" podID="0ea9a581-81d0-4255-9781-49ec01b78311" containerID="0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1" exitCode=0 Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.265057 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rm88h" event={"ID":"0ea9a581-81d0-4255-9781-49ec01b78311","Type":"ContainerDied","Data":"0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1"} Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.265065 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rm88h" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.265092 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rm88h" event={"ID":"0ea9a581-81d0-4255-9781-49ec01b78311","Type":"ContainerDied","Data":"80bd4268878a35d056fab6600789c645917917ec8ddebe16f2f25373598aeed8"} Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.265112 4848 scope.go:117] "RemoveContainer" containerID="0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.293023 4848 scope.go:117] "RemoveContainer" containerID="48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.306383 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rm88h"] Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.321729 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rm88h"] Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.331934 4848 scope.go:117] "RemoveContainer" containerID="fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.376368 4848 scope.go:117] "RemoveContainer" containerID="0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1" Dec 04 15:46:36 crc kubenswrapper[4848]: E1204 15:46:36.377364 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1\": container with ID starting with 0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1 not found: ID does not exist" containerID="0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.377401 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1"} err="failed to get container status \"0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1\": rpc error: code = NotFound desc = could not find container \"0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1\": container with ID starting with 0ffc544bf8b44794af528144be79a6d145e0f42ee977d6f265ef732e3f8304e1 not found: ID does not exist" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.377422 4848 scope.go:117] "RemoveContainer" containerID="48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1" Dec 04 15:46:36 crc kubenswrapper[4848]: E1204 15:46:36.377691 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1\": container with ID starting with 48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1 not found: ID does not exist" containerID="48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.377719 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1"} err="failed to get container status \"48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1\": rpc error: code = NotFound desc = could not find container \"48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1\": container with ID starting with 48eb56aa448279713b5d45c23b6ddba045bfd24fa7b034ee3040b7af06e36aa1 not found: ID does not exist" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.377733 4848 scope.go:117] "RemoveContainer" containerID="fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a" Dec 04 15:46:36 crc kubenswrapper[4848]: E1204 15:46:36.377971 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a\": container with ID starting with fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a not found: ID does not exist" containerID="fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.378001 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a"} err="failed to get container status \"fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a\": rpc error: code = NotFound desc = could not find container \"fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a\": container with ID starting with fc8dae71572597abe046839aec517e689faeb520bac88a0321a8aad480afe40a not found: ID does not exist" Dec 04 15:46:36 crc kubenswrapper[4848]: I1204 15:46:36.408229 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" path="/var/lib/kubelet/pods/0ea9a581-81d0-4255-9781-49ec01b78311/volumes" Dec 04 15:46:46 crc kubenswrapper[4848]: I1204 15:46:46.388132 4848 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-72f2r container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 15:46:46 crc kubenswrapper[4848]: I1204 15:46:46.388843 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72f2r" podUID="870459be-ead1-4020-8f99-ba32001241e2" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 15:46:46 crc kubenswrapper[4848]: I1204 15:46:46.596344 4848 trace.go:236] Trace[870022758]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/certified-operators-lq8mk" (04-Dec-2025 15:46:45.432) (total time: 1162ms): Dec 04 15:46:46 crc kubenswrapper[4848]: Trace[870022758]: [1.162808723s] [1.162808723s] END Dec 04 15:47:08 crc kubenswrapper[4848]: I1204 15:47:08.610582 4848 generic.go:334] "Generic (PLEG): container finished" podID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerID="ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e" exitCode=0 Dec 04 15:47:08 crc kubenswrapper[4848]: I1204 15:47:08.610677 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9k646/must-gather-vx45g" event={"ID":"cf13c537-64dd-4e6b-8de6-9181b448ebd2","Type":"ContainerDied","Data":"ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e"} Dec 04 15:47:08 crc kubenswrapper[4848]: I1204 15:47:08.611974 4848 scope.go:117] "RemoveContainer" containerID="ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e" Dec 04 15:47:09 crc kubenswrapper[4848]: I1204 15:47:09.136217 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9k646_must-gather-vx45g_cf13c537-64dd-4e6b-8de6-9181b448ebd2/gather/0.log" Dec 04 15:47:14 crc kubenswrapper[4848]: I1204 15:47:14.313744 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:47:14 crc kubenswrapper[4848]: I1204 15:47:14.315103 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:47:21 crc kubenswrapper[4848]: I1204 15:47:21.727372 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9k646/must-gather-vx45g"] Dec 04 15:47:21 crc kubenswrapper[4848]: I1204 15:47:21.728294 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-9k646/must-gather-vx45g" podUID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerName="copy" containerID="cri-o://776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987" gracePeriod=2 Dec 04 15:47:21 crc kubenswrapper[4848]: I1204 15:47:21.748409 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9k646/must-gather-vx45g"] Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.363491 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9k646_must-gather-vx45g_cf13c537-64dd-4e6b-8de6-9181b448ebd2/copy/0.log" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.364721 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.449013 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf13c537-64dd-4e6b-8de6-9181b448ebd2-must-gather-output\") pod \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\" (UID: \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\") " Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.449116 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dz5c\" (UniqueName: \"kubernetes.io/projected/cf13c537-64dd-4e6b-8de6-9181b448ebd2-kube-api-access-9dz5c\") pod \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\" (UID: \"cf13c537-64dd-4e6b-8de6-9181b448ebd2\") " Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.457538 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf13c537-64dd-4e6b-8de6-9181b448ebd2-kube-api-access-9dz5c" (OuterVolumeSpecName: "kube-api-access-9dz5c") pod "cf13c537-64dd-4e6b-8de6-9181b448ebd2" (UID: "cf13c537-64dd-4e6b-8de6-9181b448ebd2"). InnerVolumeSpecName "kube-api-access-9dz5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.552059 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dz5c\" (UniqueName: \"kubernetes.io/projected/cf13c537-64dd-4e6b-8de6-9181b448ebd2-kube-api-access-9dz5c\") on node \"crc\" DevicePath \"\"" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.634561 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf13c537-64dd-4e6b-8de6-9181b448ebd2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cf13c537-64dd-4e6b-8de6-9181b448ebd2" (UID: "cf13c537-64dd-4e6b-8de6-9181b448ebd2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.654142 4848 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cf13c537-64dd-4e6b-8de6-9181b448ebd2-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.769770 4848 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9k646_must-gather-vx45g_cf13c537-64dd-4e6b-8de6-9181b448ebd2/copy/0.log" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.770227 4848 generic.go:334] "Generic (PLEG): container finished" podID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerID="776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987" exitCode=143 Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.770288 4848 scope.go:117] "RemoveContainer" containerID="776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.770388 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9k646/must-gather-vx45g" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.800696 4848 scope.go:117] "RemoveContainer" containerID="ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.844149 4848 scope.go:117] "RemoveContainer" containerID="776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987" Dec 04 15:47:22 crc kubenswrapper[4848]: E1204 15:47:22.848108 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987\": container with ID starting with 776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987 not found: ID does not exist" containerID="776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.848153 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987"} err="failed to get container status \"776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987\": rpc error: code = NotFound desc = could not find container \"776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987\": container with ID starting with 776c8af6878c86e249ee25d6c7197c7757096662b7492c60a2dcec6c7145d987 not found: ID does not exist" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.848180 4848 scope.go:117] "RemoveContainer" containerID="ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e" Dec 04 15:47:22 crc kubenswrapper[4848]: E1204 15:47:22.857140 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e\": container with ID starting with ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e not found: ID does not exist" containerID="ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e" Dec 04 15:47:22 crc kubenswrapper[4848]: I1204 15:47:22.857193 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e"} err="failed to get container status \"ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e\": rpc error: code = NotFound desc = could not find container \"ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e\": container with ID starting with ed5ead0610ce381d4513e604c9c4b063e33c46fcf80e7f85862db3935ce54a7e not found: ID does not exist" Dec 04 15:47:24 crc kubenswrapper[4848]: I1204 15:47:24.408057 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" path="/var/lib/kubelet/pods/cf13c537-64dd-4e6b-8de6-9181b448ebd2/volumes" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.938196 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9vlcq"] Dec 04 15:47:40 crc kubenswrapper[4848]: E1204 15:47:40.939834 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerName="gather" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.939858 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerName="gather" Dec 04 15:47:40 crc kubenswrapper[4848]: E1204 15:47:40.939889 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" containerName="extract-utilities" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.939897 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" containerName="extract-utilities" Dec 04 15:47:40 crc kubenswrapper[4848]: E1204 15:47:40.939913 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" containerName="extract-content" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.939921 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" containerName="extract-content" Dec 04 15:47:40 crc kubenswrapper[4848]: E1204 15:47:40.939961 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" containerName="registry-server" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.939971 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" containerName="registry-server" Dec 04 15:47:40 crc kubenswrapper[4848]: E1204 15:47:40.939992 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerName="copy" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.939998 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerName="copy" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.940344 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerName="gather" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.940367 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf13c537-64dd-4e6b-8de6-9181b448ebd2" containerName="copy" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.940391 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ea9a581-81d0-4255-9781-49ec01b78311" containerName="registry-server" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.942619 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:40 crc kubenswrapper[4848]: I1204 15:47:40.972098 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9vlcq"] Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.102668 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-catalog-content\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.103071 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mshc\" (UniqueName: \"kubernetes.io/projected/4190c580-bd4e-49d0-8c58-cb27c7490e14-kube-api-access-9mshc\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.103258 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-utilities\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.205224 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-catalog-content\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.205289 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mshc\" (UniqueName: \"kubernetes.io/projected/4190c580-bd4e-49d0-8c58-cb27c7490e14-kube-api-access-9mshc\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.205413 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-utilities\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.206107 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-utilities\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.206378 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-catalog-content\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.234824 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mshc\" (UniqueName: \"kubernetes.io/projected/4190c580-bd4e-49d0-8c58-cb27c7490e14-kube-api-access-9mshc\") pod \"community-operators-9vlcq\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.268130 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:41 crc kubenswrapper[4848]: I1204 15:47:41.836420 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9vlcq"] Dec 04 15:47:42 crc kubenswrapper[4848]: I1204 15:47:42.008857 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vlcq" event={"ID":"4190c580-bd4e-49d0-8c58-cb27c7490e14","Type":"ContainerStarted","Data":"86eecd7599bc79be682d6ed1cff4a0b5d88a47ce5abddabbf5a1317d099f9499"} Dec 04 15:47:42 crc kubenswrapper[4848]: I1204 15:47:42.992811 4848 scope.go:117] "RemoveContainer" containerID="3982eba4321306a24b0b6c4d7ce674cfb14531ff094cf235878a118c24b3da90" Dec 04 15:47:43 crc kubenswrapper[4848]: I1204 15:47:43.015967 4848 scope.go:117] "RemoveContainer" containerID="16adfe1c2595a3dea3f34e4665f00ffa325bbc3cf5d1fc04e3618212fd17d281" Dec 04 15:47:43 crc kubenswrapper[4848]: I1204 15:47:43.021326 4848 generic.go:334] "Generic (PLEG): container finished" podID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerID="c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68" exitCode=0 Dec 04 15:47:43 crc kubenswrapper[4848]: I1204 15:47:43.021369 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vlcq" event={"ID":"4190c580-bd4e-49d0-8c58-cb27c7490e14","Type":"ContainerDied","Data":"c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68"} Dec 04 15:47:44 crc kubenswrapper[4848]: I1204 15:47:44.313843 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:47:44 crc kubenswrapper[4848]: I1204 15:47:44.314207 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:47:45 crc kubenswrapper[4848]: I1204 15:47:45.042298 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vlcq" event={"ID":"4190c580-bd4e-49d0-8c58-cb27c7490e14","Type":"ContainerStarted","Data":"a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696"} Dec 04 15:47:46 crc kubenswrapper[4848]: I1204 15:47:46.054110 4848 generic.go:334] "Generic (PLEG): container finished" podID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerID="a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696" exitCode=0 Dec 04 15:47:46 crc kubenswrapper[4848]: I1204 15:47:46.054153 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vlcq" event={"ID":"4190c580-bd4e-49d0-8c58-cb27c7490e14","Type":"ContainerDied","Data":"a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696"} Dec 04 15:47:47 crc kubenswrapper[4848]: I1204 15:47:47.067022 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vlcq" event={"ID":"4190c580-bd4e-49d0-8c58-cb27c7490e14","Type":"ContainerStarted","Data":"05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9"} Dec 04 15:47:47 crc kubenswrapper[4848]: I1204 15:47:47.092666 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9vlcq" podStartSLOduration=3.655668355 podStartE2EDuration="7.092647258s" podCreationTimestamp="2025-12-04 15:47:40 +0000 UTC" firstStartedPulling="2025-12-04 15:47:43.023040169 +0000 UTC m=+7166.965536697" lastFinishedPulling="2025-12-04 15:47:46.460019072 +0000 UTC m=+7170.402515600" observedRunningTime="2025-12-04 15:47:47.082865878 +0000 UTC m=+7171.025362406" watchObservedRunningTime="2025-12-04 15:47:47.092647258 +0000 UTC m=+7171.035143786" Dec 04 15:47:51 crc kubenswrapper[4848]: I1204 15:47:51.268978 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:51 crc kubenswrapper[4848]: I1204 15:47:51.269617 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:51 crc kubenswrapper[4848]: I1204 15:47:51.318581 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:52 crc kubenswrapper[4848]: I1204 15:47:52.172729 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:52 crc kubenswrapper[4848]: I1204 15:47:52.225459 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9vlcq"] Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.144179 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9vlcq" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerName="registry-server" containerID="cri-o://05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9" gracePeriod=2 Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.709768 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.833807 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mshc\" (UniqueName: \"kubernetes.io/projected/4190c580-bd4e-49d0-8c58-cb27c7490e14-kube-api-access-9mshc\") pod \"4190c580-bd4e-49d0-8c58-cb27c7490e14\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.834104 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-catalog-content\") pod \"4190c580-bd4e-49d0-8c58-cb27c7490e14\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.834263 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-utilities\") pod \"4190c580-bd4e-49d0-8c58-cb27c7490e14\" (UID: \"4190c580-bd4e-49d0-8c58-cb27c7490e14\") " Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.835680 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-utilities" (OuterVolumeSpecName: "utilities") pod "4190c580-bd4e-49d0-8c58-cb27c7490e14" (UID: "4190c580-bd4e-49d0-8c58-cb27c7490e14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.842059 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4190c580-bd4e-49d0-8c58-cb27c7490e14-kube-api-access-9mshc" (OuterVolumeSpecName: "kube-api-access-9mshc") pod "4190c580-bd4e-49d0-8c58-cb27c7490e14" (UID: "4190c580-bd4e-49d0-8c58-cb27c7490e14"). InnerVolumeSpecName "kube-api-access-9mshc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.894800 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4190c580-bd4e-49d0-8c58-cb27c7490e14" (UID: "4190c580-bd4e-49d0-8c58-cb27c7490e14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.937200 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.937247 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4190c580-bd4e-49d0-8c58-cb27c7490e14-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:47:54 crc kubenswrapper[4848]: I1204 15:47:54.937266 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mshc\" (UniqueName: \"kubernetes.io/projected/4190c580-bd4e-49d0-8c58-cb27c7490e14-kube-api-access-9mshc\") on node \"crc\" DevicePath \"\"" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.155656 4848 generic.go:334] "Generic (PLEG): container finished" podID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerID="05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9" exitCode=0 Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.155709 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vlcq" event={"ID":"4190c580-bd4e-49d0-8c58-cb27c7490e14","Type":"ContainerDied","Data":"05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9"} Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.155715 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9vlcq" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.155751 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9vlcq" event={"ID":"4190c580-bd4e-49d0-8c58-cb27c7490e14","Type":"ContainerDied","Data":"86eecd7599bc79be682d6ed1cff4a0b5d88a47ce5abddabbf5a1317d099f9499"} Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.155781 4848 scope.go:117] "RemoveContainer" containerID="05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.180254 4848 scope.go:117] "RemoveContainer" containerID="a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.202182 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9vlcq"] Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.214489 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9vlcq"] Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.225271 4848 scope.go:117] "RemoveContainer" containerID="c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.269704 4848 scope.go:117] "RemoveContainer" containerID="05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9" Dec 04 15:47:55 crc kubenswrapper[4848]: E1204 15:47:55.276761 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9\": container with ID starting with 05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9 not found: ID does not exist" containerID="05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.276848 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9"} err="failed to get container status \"05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9\": rpc error: code = NotFound desc = could not find container \"05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9\": container with ID starting with 05aaae8ddf08541982748895f2407201ccc49f7615962ffa1f9669ae28f876a9 not found: ID does not exist" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.276900 4848 scope.go:117] "RemoveContainer" containerID="a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696" Dec 04 15:47:55 crc kubenswrapper[4848]: E1204 15:47:55.277744 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696\": container with ID starting with a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696 not found: ID does not exist" containerID="a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.277787 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696"} err="failed to get container status \"a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696\": rpc error: code = NotFound desc = could not find container \"a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696\": container with ID starting with a6766d58b0407438529ab86a4ca57b714f91bc00488ad75f4e294567286c4696 not found: ID does not exist" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.277841 4848 scope.go:117] "RemoveContainer" containerID="c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68" Dec 04 15:47:55 crc kubenswrapper[4848]: E1204 15:47:55.278212 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68\": container with ID starting with c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68 not found: ID does not exist" containerID="c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68" Dec 04 15:47:55 crc kubenswrapper[4848]: I1204 15:47:55.278257 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68"} err="failed to get container status \"c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68\": rpc error: code = NotFound desc = could not find container \"c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68\": container with ID starting with c801c46fffc439998127d96f86d5aa233328c11ace1e4287f830398c85e6ff68 not found: ID does not exist" Dec 04 15:47:56 crc kubenswrapper[4848]: I1204 15:47:56.413128 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" path="/var/lib/kubelet/pods/4190c580-bd4e-49d0-8c58-cb27c7490e14/volumes" Dec 04 15:48:14 crc kubenswrapper[4848]: I1204 15:48:14.314992 4848 patch_prober.go:28] interesting pod/machine-config-daemon-jspn5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 15:48:14 crc kubenswrapper[4848]: I1204 15:48:14.315768 4848 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 15:48:14 crc kubenswrapper[4848]: I1204 15:48:14.315835 4848 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" Dec 04 15:48:14 crc kubenswrapper[4848]: I1204 15:48:14.317080 4848 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d"} pod="openshift-machine-config-operator/machine-config-daemon-jspn5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 15:48:14 crc kubenswrapper[4848]: I1204 15:48:14.317219 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" containerName="machine-config-daemon" containerID="cri-o://b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" gracePeriod=600 Dec 04 15:48:15 crc kubenswrapper[4848]: I1204 15:48:15.552271 4848 patch_prober.go:28] interesting pod/console-operator-58897d9998-2cmwc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 15:48:15 crc kubenswrapper[4848]: I1204 15:48:15.552709 4848 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-2cmwc" podUID="e52ecea7-0826-4ceb-8d32-f21f4c3fd181" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 15:48:16 crc kubenswrapper[4848]: E1204 15:48:16.112616 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:48:16 crc kubenswrapper[4848]: I1204 15:48:16.404302 4848 generic.go:334] "Generic (PLEG): container finished" podID="bedd115c-06c2-484b-88ab-4748c86b4531" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" exitCode=0 Dec 04 15:48:16 crc kubenswrapper[4848]: I1204 15:48:16.406498 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" event={"ID":"bedd115c-06c2-484b-88ab-4748c86b4531","Type":"ContainerDied","Data":"b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d"} Dec 04 15:48:16 crc kubenswrapper[4848]: I1204 15:48:16.406542 4848 scope.go:117] "RemoveContainer" containerID="78749c589f78ef3cb53b6338248052af0028bcb29d8a4feb7475ba80de9997d6" Dec 04 15:48:16 crc kubenswrapper[4848]: I1204 15:48:16.408568 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:48:16 crc kubenswrapper[4848]: E1204 15:48:16.409202 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:48:31 crc kubenswrapper[4848]: I1204 15:48:31.394450 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:48:31 crc kubenswrapper[4848]: E1204 15:48:31.396549 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:48:44 crc kubenswrapper[4848]: I1204 15:48:44.393942 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:48:44 crc kubenswrapper[4848]: E1204 15:48:44.394644 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:48:57 crc kubenswrapper[4848]: I1204 15:48:57.393477 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:48:57 crc kubenswrapper[4848]: E1204 15:48:57.394390 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:49:12 crc kubenswrapper[4848]: I1204 15:49:12.393232 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:49:12 crc kubenswrapper[4848]: E1204 15:49:12.394002 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:49:25 crc kubenswrapper[4848]: I1204 15:49:25.393798 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:49:25 crc kubenswrapper[4848]: E1204 15:49:25.395266 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:49:40 crc kubenswrapper[4848]: I1204 15:49:40.394001 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:49:40 crc kubenswrapper[4848]: E1204 15:49:40.395055 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:49:53 crc kubenswrapper[4848]: I1204 15:49:53.394272 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:49:53 crc kubenswrapper[4848]: E1204 15:49:53.395391 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:50:08 crc kubenswrapper[4848]: I1204 15:50:08.394561 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:50:08 crc kubenswrapper[4848]: E1204 15:50:08.395604 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.320998 4848 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kf5bj"] Dec 04 15:50:20 crc kubenswrapper[4848]: E1204 15:50:20.322394 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerName="registry-server" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.322419 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerName="registry-server" Dec 04 15:50:20 crc kubenswrapper[4848]: E1204 15:50:20.322451 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerName="extract-utilities" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.322463 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerName="extract-utilities" Dec 04 15:50:20 crc kubenswrapper[4848]: E1204 15:50:20.322505 4848 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerName="extract-content" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.322517 4848 state_mem.go:107] "Deleted CPUSet assignment" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerName="extract-content" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.323013 4848 memory_manager.go:354] "RemoveStaleState removing state" podUID="4190c580-bd4e-49d0-8c58-cb27c7490e14" containerName="registry-server" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.325722 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.342449 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kf5bj"] Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.394368 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:50:20 crc kubenswrapper[4848]: E1204 15:50:20.395071 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.446143 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-catalog-content\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.446305 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbth4\" (UniqueName: \"kubernetes.io/projected/48451b38-47c5-44ef-aecb-b4c5cd1db769-kube-api-access-lbth4\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.446399 4848 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-utilities\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.549701 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-utilities\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.549844 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-catalog-content\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.549962 4848 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbth4\" (UniqueName: \"kubernetes.io/projected/48451b38-47c5-44ef-aecb-b4c5cd1db769-kube-api-access-lbth4\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.550223 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-utilities\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.550716 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-catalog-content\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.570542 4848 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbth4\" (UniqueName: \"kubernetes.io/projected/48451b38-47c5-44ef-aecb-b4c5cd1db769-kube-api-access-lbth4\") pod \"redhat-operators-kf5bj\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:20 crc kubenswrapper[4848]: I1204 15:50:20.649776 4848 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:21 crc kubenswrapper[4848]: I1204 15:50:21.158136 4848 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kf5bj"] Dec 04 15:50:21 crc kubenswrapper[4848]: I1204 15:50:21.788355 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kf5bj" event={"ID":"48451b38-47c5-44ef-aecb-b4c5cd1db769","Type":"ContainerStarted","Data":"4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a"} Dec 04 15:50:21 crc kubenswrapper[4848]: I1204 15:50:21.788636 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kf5bj" event={"ID":"48451b38-47c5-44ef-aecb-b4c5cd1db769","Type":"ContainerStarted","Data":"68d985a17eaef1b6973cc979c00c41a274cf559ec4d4cff60a772f987df7d5a0"} Dec 04 15:50:22 crc kubenswrapper[4848]: I1204 15:50:22.807473 4848 generic.go:334] "Generic (PLEG): container finished" podID="48451b38-47c5-44ef-aecb-b4c5cd1db769" containerID="4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a" exitCode=0 Dec 04 15:50:22 crc kubenswrapper[4848]: I1204 15:50:22.807653 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kf5bj" event={"ID":"48451b38-47c5-44ef-aecb-b4c5cd1db769","Type":"ContainerDied","Data":"4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a"} Dec 04 15:50:22 crc kubenswrapper[4848]: I1204 15:50:22.808160 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kf5bj" event={"ID":"48451b38-47c5-44ef-aecb-b4c5cd1db769","Type":"ContainerStarted","Data":"d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4"} Dec 04 15:50:26 crc kubenswrapper[4848]: I1204 15:50:26.848157 4848 generic.go:334] "Generic (PLEG): container finished" podID="48451b38-47c5-44ef-aecb-b4c5cd1db769" containerID="d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4" exitCode=0 Dec 04 15:50:26 crc kubenswrapper[4848]: I1204 15:50:26.848231 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kf5bj" event={"ID":"48451b38-47c5-44ef-aecb-b4c5cd1db769","Type":"ContainerDied","Data":"d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4"} Dec 04 15:50:27 crc kubenswrapper[4848]: I1204 15:50:27.866389 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kf5bj" event={"ID":"48451b38-47c5-44ef-aecb-b4c5cd1db769","Type":"ContainerStarted","Data":"22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869"} Dec 04 15:50:27 crc kubenswrapper[4848]: I1204 15:50:27.889631 4848 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kf5bj" podStartSLOduration=2.334176831 podStartE2EDuration="7.889610657s" podCreationTimestamp="2025-12-04 15:50:20 +0000 UTC" firstStartedPulling="2025-12-04 15:50:21.790581116 +0000 UTC m=+7325.733077664" lastFinishedPulling="2025-12-04 15:50:27.346014962 +0000 UTC m=+7331.288511490" observedRunningTime="2025-12-04 15:50:27.886332566 +0000 UTC m=+7331.828829124" watchObservedRunningTime="2025-12-04 15:50:27.889610657 +0000 UTC m=+7331.832107185" Dec 04 15:50:30 crc kubenswrapper[4848]: I1204 15:50:30.650371 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:30 crc kubenswrapper[4848]: I1204 15:50:30.650713 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:31 crc kubenswrapper[4848]: I1204 15:50:31.770187 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kf5bj" podUID="48451b38-47c5-44ef-aecb-b4c5cd1db769" containerName="registry-server" probeResult="failure" output=< Dec 04 15:50:31 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:50:31 crc kubenswrapper[4848]: > Dec 04 15:50:33 crc kubenswrapper[4848]: I1204 15:50:33.393629 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:50:33 crc kubenswrapper[4848]: E1204 15:50:33.394336 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:50:41 crc kubenswrapper[4848]: I1204 15:50:41.702075 4848 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kf5bj" podUID="48451b38-47c5-44ef-aecb-b4c5cd1db769" containerName="registry-server" probeResult="failure" output=< Dec 04 15:50:41 crc kubenswrapper[4848]: timeout: failed to connect service ":50051" within 1s Dec 04 15:50:41 crc kubenswrapper[4848]: > Dec 04 15:50:45 crc kubenswrapper[4848]: I1204 15:50:45.393246 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:50:45 crc kubenswrapper[4848]: E1204 15:50:45.394104 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:50:50 crc kubenswrapper[4848]: I1204 15:50:50.698698 4848 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:50 crc kubenswrapper[4848]: I1204 15:50:50.748748 4848 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:51 crc kubenswrapper[4848]: I1204 15:50:51.517432 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kf5bj"] Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.177727 4848 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kf5bj" podUID="48451b38-47c5-44ef-aecb-b4c5cd1db769" containerName="registry-server" containerID="cri-o://22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869" gracePeriod=2 Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.760543 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.859352 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbth4\" (UniqueName: \"kubernetes.io/projected/48451b38-47c5-44ef-aecb-b4c5cd1db769-kube-api-access-lbth4\") pod \"48451b38-47c5-44ef-aecb-b4c5cd1db769\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.860212 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-catalog-content\") pod \"48451b38-47c5-44ef-aecb-b4c5cd1db769\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.860252 4848 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-utilities\") pod \"48451b38-47c5-44ef-aecb-b4c5cd1db769\" (UID: \"48451b38-47c5-44ef-aecb-b4c5cd1db769\") " Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.861239 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-utilities" (OuterVolumeSpecName: "utilities") pod "48451b38-47c5-44ef-aecb-b4c5cd1db769" (UID: "48451b38-47c5-44ef-aecb-b4c5cd1db769"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.872244 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48451b38-47c5-44ef-aecb-b4c5cd1db769-kube-api-access-lbth4" (OuterVolumeSpecName: "kube-api-access-lbth4") pod "48451b38-47c5-44ef-aecb-b4c5cd1db769" (UID: "48451b38-47c5-44ef-aecb-b4c5cd1db769"). InnerVolumeSpecName "kube-api-access-lbth4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.963397 4848 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.963432 4848 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbth4\" (UniqueName: \"kubernetes.io/projected/48451b38-47c5-44ef-aecb-b4c5cd1db769-kube-api-access-lbth4\") on node \"crc\" DevicePath \"\"" Dec 04 15:50:52 crc kubenswrapper[4848]: I1204 15:50:52.963874 4848 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48451b38-47c5-44ef-aecb-b4c5cd1db769" (UID: "48451b38-47c5-44ef-aecb-b4c5cd1db769"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.065650 4848 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48451b38-47c5-44ef-aecb-b4c5cd1db769-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.191377 4848 generic.go:334] "Generic (PLEG): container finished" podID="48451b38-47c5-44ef-aecb-b4c5cd1db769" containerID="22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869" exitCode=0 Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.191431 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kf5bj" event={"ID":"48451b38-47c5-44ef-aecb-b4c5cd1db769","Type":"ContainerDied","Data":"22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869"} Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.191463 4848 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kf5bj" event={"ID":"48451b38-47c5-44ef-aecb-b4c5cd1db769","Type":"ContainerDied","Data":"68d985a17eaef1b6973cc979c00c41a274cf559ec4d4cff60a772f987df7d5a0"} Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.191489 4848 scope.go:117] "RemoveContainer" containerID="22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.191484 4848 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kf5bj" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.217908 4848 scope.go:117] "RemoveContainer" containerID="d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.233045 4848 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kf5bj"] Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.246006 4848 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kf5bj"] Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.249026 4848 scope.go:117] "RemoveContainer" containerID="4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.324363 4848 scope.go:117] "RemoveContainer" containerID="22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869" Dec 04 15:50:53 crc kubenswrapper[4848]: E1204 15:50:53.325871 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869\": container with ID starting with 22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869 not found: ID does not exist" containerID="22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.325903 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869"} err="failed to get container status \"22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869\": rpc error: code = NotFound desc = could not find container \"22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869\": container with ID starting with 22f752491aa717102d988e29c013aea5e502e14a30d0ac53ab979343f2326869 not found: ID does not exist" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.325928 4848 scope.go:117] "RemoveContainer" containerID="d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4" Dec 04 15:50:53 crc kubenswrapper[4848]: E1204 15:50:53.326332 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4\": container with ID starting with d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4 not found: ID does not exist" containerID="d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.326348 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4"} err="failed to get container status \"d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4\": rpc error: code = NotFound desc = could not find container \"d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4\": container with ID starting with d89e7099a46518352ddb17875fafbe5723c2d6b39e760a88e1fcc3690d0723a4 not found: ID does not exist" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.326364 4848 scope.go:117] "RemoveContainer" containerID="4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a" Dec 04 15:50:53 crc kubenswrapper[4848]: E1204 15:50:53.326561 4848 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a\": container with ID starting with 4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a not found: ID does not exist" containerID="4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a" Dec 04 15:50:53 crc kubenswrapper[4848]: I1204 15:50:53.326582 4848 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a"} err="failed to get container status \"4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a\": rpc error: code = NotFound desc = could not find container \"4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a\": container with ID starting with 4717702e9904f2d7f30c435c95e562e8eeafa19cbcc4365dd7d7e15c56fe575a not found: ID does not exist" Dec 04 15:50:54 crc kubenswrapper[4848]: I1204 15:50:54.408269 4848 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48451b38-47c5-44ef-aecb-b4c5cd1db769" path="/var/lib/kubelet/pods/48451b38-47c5-44ef-aecb-b4c5cd1db769/volumes" Dec 04 15:50:56 crc kubenswrapper[4848]: I1204 15:50:56.405006 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:50:56 crc kubenswrapper[4848]: E1204 15:50:56.405682 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" Dec 04 15:51:09 crc kubenswrapper[4848]: I1204 15:51:09.394454 4848 scope.go:117] "RemoveContainer" containerID="b0df3dd043663a31f36355a864bdd539d0acfdf60cb4aadbe530c03cc67dd67d" Dec 04 15:51:09 crc kubenswrapper[4848]: E1204 15:51:09.395345 4848 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jspn5_openshift-machine-config-operator(bedd115c-06c2-484b-88ab-4748c86b4531)\"" pod="openshift-machine-config-operator/machine-config-daemon-jspn5" podUID="bedd115c-06c2-484b-88ab-4748c86b4531" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114326774024457 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114326775017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114310102016471 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114310102015441 5ustar corecore